source stringlengths 31 227 | text stringlengths 9 2k |
|---|---|
https://en.wikipedia.org/wiki/Bot%C3%A1nica | A botánica (often written botanica and less commonly known as a hierbería or botica) is a religious goods store. The name botánica is Spanish and translates as "botany" or "plant store," referring to these establishments' function as dispensaries of medicinal herbs. Botánicas are common in many Hispanic American countries and Latino communities around the world. Such establishments sell folk medicine, herbs, candles, and statues for Saints and popular gods. They also carry oils, incense, perfumes, and books. Such stores have become increasingly popular in the United States as the Latino communities they serve have grown in that country. A botánica is a site of healing and support, such that one owner says they are a "place of mysteries" due to the metaphysical appreciation of mystery as a synonym for spirit and divinity.
Most botánicas sell products and services associated with spiritual practices such as Candomblé, Curanderismo, Espiritismo, Macumba, Palo Monte, Santa Muerte, María Lionza and Santería. Whether these items are viewed as cultural imports or adaptive responses on the part of immigrants to a new social environment, the majority of these products and services are used by those who seek guidance in their spiritual and social lives. Botánicas provide their patrons with access to power: power from the natural world, the social world, and the world of the spirits. Devotees, in turn, use this power to meet the challenges of ordinary life: problems of health, wealth, and love. People come to the botánica with a host of struggles and problems, and the botánica offers hope from these troubles.
There is extensive research and literature on botánicas as dispensers of healthcare in the Latino communities of the United States. Without access to professional health services, many Latinos have found effective care in the herbal treatments and psychological support that botánicas offer.
History
Botánicas extend centuries-old practices of using plants and herbs to |
https://en.wikipedia.org/wiki/Myriocin | Myriocin, also known as antibiotic ISP-1 and thermozymocidin, is a non-proteinogenic amino acid derived from certain thermophilic fungi.
Myriocin is a very potent inhibitor of serine palmitoyltransferase, the first step in sphingosine biosynthesis. Due to this property, it is used in biochemical research as a tool for depleting cells of sphingolipids.
Myriocin was shown to inhibit the proliferation of an IL-2-dependent mouse cytotoxic T cell line.
Myriocin possesses immunosuppressant activity. It is reported to be 10- to 100-fold more potent than ciclosporin.
The multiple sclerosis drug fingolimod was derived from myriocin by using structure–activity relationship studies to determine the parts of the molecule important to its activity. |
https://en.wikipedia.org/wiki/Proper%20acceleration | In relativity theory, proper acceleration is the physical acceleration (i.e., measurable acceleration as by an accelerometer) experienced by an object. It is thus acceleration relative to a free-fall, or inertial, observer who is momentarily at rest relative to the object being measured. Gravitation therefore does not cause proper acceleration, because the same gravity acts equally on the inertial observer. As a consequence, all inertial observers always have a proper acceleration of zero.
Proper acceleration contrasts with coordinate acceleration, which is dependent on choice of coordinate systems and thus upon choice of observers (see three-acceleration in special relativity).
In the standard inertial coordinates of special relativity, for unidirectional motion, proper acceleration is the rate of change of proper velocity with respect to coordinate time.
In an inertial frame in which the object is momentarily at rest, the proper acceleration 3-vector, combined with a zero time-component, yields the object's four-acceleration, which makes proper-acceleration's magnitude Lorentz-invariant. Thus the concept is useful: (i) with accelerated coordinate systems, (ii) at relativistic speeds, and (iii) in curved spacetime.
In an accelerating rocket after launch, or even in a rocket standing on the launch pad, the proper acceleration is the acceleration felt by the occupants, and which is described as g-force (which is not a force but rather an acceleration; see that article for more discussion) delivered by the vehicle only. The "acceleration of gravity" (involved in the "force of gravity") never contributes to proper acceleration in any circumstances, and thus the proper acceleration felt by observers standing on the ground is due to the mechanical force from the ground, not due to the "force" or "acceleration" of gravity. If the ground is removed and the observer allowed to free-fall, the observer will experience coordinate acceleration, but no proper acceleration, a |
https://en.wikipedia.org/wiki/Electronic%20circuit%20design | Electronic circuit design comprises the analysis and synthesis of electronic circuits.
Methods
To design any electrical circuit, either analog or digital, electrical engineers need to be able to predict the voltages and currents at all places within the circuit. Linear circuits, that is, circuits wherein the outputs are linearly dependent on the inputs, can be analyzed by hand using complex analysis. Simple nonlinear circuits can also be analyzed in this way. Specialized software has been created to analyze circuits that are either too complicated or too nonlinear to analyze by hand.
Circuit simulation software allows engineers to design circuits more efficiently, reducing the time cost and risk of error involved in building circuit prototypes. Some of these make use of hardware description languages such as VHDL or Verilog.
Network simulation software
More complex circuits are analyzed with circuit simulation software such as SPICE and EMTP.
Linearization around operating point
When faced with a new circuit, the software first tries to find a steady state solution wherein all the nodes conform to Kirchhoff's Current Law and the voltages across and through each element of the circuit conform to the voltage/current equations governing that element.
Once the steady state solution is found, the software can analyze the response to perturbations using piecewise approximation, harmonic balance or other methods.
Piece-wise linear approximation
Software such as the PLECS interface to Simulink uses piecewise linear approximation of the equations governing the elements of a circuit. The circuit is treated as a completely linear network of ideal diodes. Every time a diode switches from on to off or vice versa, the configuration of the linear network changes. Adding more detail to the approximation of equations increases the accuracy of the simulation, but also increases its running time.
Synthesis
Simple circuits may be designed by connecting a number of elements or |
https://en.wikipedia.org/wiki/Yuri%20Burago | Yuri Dmitrievich Burago (; born 1936) is a Russian mathematician. He works in differential and convex geometry.
Education and career
Burago studied at Leningrad University, where he obtained his Ph.D. and Habilitation degrees. His advisors were Victor Zalgaller and Aleksandr Aleksandrov.
Yuri is a creator (with his students Perelman and Petrunin, and M. Gromov) of what is known now as Alexandrov Geometry. Also brought geometric inequalities to the state of art.
Burago is the head of the Laboratory of Geometry and Topology that is part of the St. Petersburg Department of Steklov Institute of Mathematics. He took part in a report for the United States Civilian Research and Development Foundation for the Independent States of the former Soviet Union.
Works
His other books and papers include:
Geometry III: Theory of Surfaces (1992)
Potential Theory and Function Theory for Irregular Regions (1969)
Isoperimetric inequalities in the theory of surfaces of bounded external curvature (1970)
Students
He has advised Grigori Perelman, who solved the Poincaré conjecture, one of the seven Millennium Prize Problems. Burago was an advisor to Perelman during the latter's post-graduate research at St. Petersburg Department of Steklov Institute of Mathematics.
Footnotes
External links
Burago's page on the site of Steklov Mathematical Institute
Yuri Dmitrievich Burago in the Oberwolfach Photo Collection
Soviet mathematicians
Geometers
Differential geometers
1936 births
Living people
Saint Petersburg State University alumni
20th-century Russian mathematicians
21st-century Russian mathematicians |
https://en.wikipedia.org/wiki/Niels%20Bohr%20International%20Gold%20Medal | The Niels Bohr International Gold Medal is an international engineering award. It has been awarded since 1955 for "outstanding work by an engineer or physicist for the peaceful utilization of atomic energy". The medal is administered by the Danish Society of Engineers (Denmark) in collaboration with the Niels Bohr Institute and the Royal Danish Academy of Sciences. It was awarded 10 times between 1955 and 1982 and again in 2013. The first recipient was Niels Bohr himself who received the medal in connection with his 70th birthday.
2013 laureate
Alain Aspect, regarded as an outstanding figure in optical and atomic physics, was awarded the medal for his experiments on the Bell's inequalities test. It was presented on 7 October 2013 by Queen Margrethe and Prince Henrik at a special event at the Honorary Residence in the Carlsberg Academy.
Recipients
The following scientists have been awarded the Niels Bohr Medal:
Niels Bohr, 1955
John Cockcroft, 1958
George de Hevesy, 1961
Pyotr Kapitsa, 1965
Isidor Isaac Rabi, 1967
Werner Karl Heisenberg, 1970
Richard P. Feynman, 1973
Hans A. Bethe, 1976
Charles H. Townes, 1979
John Archibald Wheeler, 1982
Alain Aspect, 2013
Jens Nørskov, 2018
Ewine van Dishoeck, 2022
See also
UNESCO Niels Bohr Medal
List of engineering awards
List of physics awards |
https://en.wikipedia.org/wiki/Elmer%20A.%20Sperry%20Award | The Elmer A. Sperry Award, named after the inventor and entrepreneur, is an American transportation engineering prize.
It has been given since 1955 for "a distinguished engineering contribution which, through application, proved in actual service, has advanced the art of transportation whether by land, sea, air, or space." The prize is given jointly by the American Institute of Aeronautics and Astronautics, Institute of Electrical and Electronics Engineers, Society of Automotive Engineers, Society of Naval Architects and Marine Engineers, American Society of Civil Engineers, and the American Society of Mechanical Engineers (which administers it). The purpose of the award is to encourage progress in the engineering of transportation.
Recipients
Source: Elmer A. Sperry award
1955 William Francis Gibbs, for the development of the SS United States
1956 Donald W. Douglas, for the DC series of air transport planes
1957 Harold L. Hamilton, Richard M. Dilworth and Eugene W. Kettering, for developing the diesel-electric locomotive
1958 Ferdinand Porsche (in memoriam) and Heinz Nordhoff, for development of the Volkswagen automobile
1959 Sir Geoffrey De Havilland, Major Frank Halford (in memoriam) and Charles C. Walker, for the first jet-powered passenger aircraft and engines
1960 Frederick Darcy Braddon, Sperry Gyroscope Company, for the three-axis gyroscopic navigational reference
1961 Robert Gilmore LeTourneau, Firestone Tire and Rubber Company, for large capacity earth moving equipment and giant size tires
1962 Lloyd J. Hibbard, for applying the ignitron rectifier to railroad motive power
1963 Earl A. Thompson, for design and development of the first successful automatic automobile transmission
1964 Igor Sikorsky and Michael E. Gluhareff, Sikorsky Aircraft Division, United Aircraft Corporation, for developing the high-lift helicopter leading to the Skycrane
1965 Maynard Pennell, Richard L. Rouzie, John E. Steiner, William H. Cook and Richard L. Loesch, Jr., Commercial A |
https://en.wikipedia.org/wiki/Hoover%20Medal | The Hoover Medal is an American engineering prize.
It has been given since 1930 for "outstanding extra-career services by engineers to humanity". The prize is given jointly by the American Institute of Chemical Engineers, American Institute of Mining, Metallurgical, and Petroleum Engineers, American Society of Civil Engineers, Institute of Electrical and Electronics Engineers, and American Society of Mechanical Engineers (ASME), which administers it. It is named for Herbert Hoover, the first recipient, who was an engineer by profession.
Past recipients
Source:ASME
See also
List of engineering awards
List of mechanical engineering awards
List of awards for contributions to society
List of awards named after people |
https://en.wikipedia.org/wiki/Posterior%20atlantoaxial%20ligament | The posterior atlantoaxial ligament is a broad, thin membrane attached, above, to the lower border of the posterior arch of the atlas; below, to the upper edges of the laminæ of the axis.
It supplies the place of the ligamenta flava, and is in relation, behind, with the Obliqui capitis inferiores.
See also
Atlanto-axial joint |
https://en.wikipedia.org/wiki/Anterior%20atlantoaxial%20ligament | The anterior atlantoaxial ligament is a strong membrane, fixed above the lower border of the anterior arch of the atlas; below, to the front of the body of the axis.
It is strengthened in the middle line by a rounded cord, which connects the tubercle on the anterior arch of the atlas to the body of the axis. It is a continuation upward of the anterior longitudinal ligament.
Anatomy
Anatomical relations
The anterior atlantoaxial ligament is situated anterior to the longus capitis muscle.
See also
Atlanto-axial joint |
https://en.wikipedia.org/wiki/Anterior%20atlantooccipital%20membrane | The anterior atlantooccipital membrane (anterior atlantooccipital ligament) is a broad, dense membrane extending between the anterior margin of the foramen magnum (superiorly), and (the superior margin of) the anterior arch of atlas (inferiorly).
The membrane helps limit excessive movement at the atlanto-occipital joints.
Anatomy
Structure
It is composed of broad, densely woven fibers; especially towards the midline where the membrane is continuous medially with the anterior longitudinal ligament. It is innervated by the cervical spinal nerve 1.
Relations
Medially, it is continuous with the anterior longitudinal ligament.
Laterally, it is blends with either articular capsule.
This membrane is in relation in anteriorly with the rectus capitis anterior muscles, and posteriorly with the alar ligaments.
See also
Posterior atlantooccipital membrane |
https://en.wikipedia.org/wiki/Anterior%20auricular%20branches | The anterior auricular branches of the superficial temporal artery are distributed to the anterior portion of the auricula, the lobule, and part of the external meatus, anastomosing with the posterior auricular. They supply the external acoustic meatus and the visible part of the ear.
Arterial Vascular Pattern of the Auricle:
"The superior and the inferior anterior auricular artery provided the vascular supply to the helical rim, forming an arcade, i.e. helical rim arcade. On the superior third of the helical rim another arcade was confirmed between the superior anterior auricular artery and the posterior auricular artery (PAA), i.e. the helical arcade. The perforators of the PAA were identified lying in a vertical line 1 cm posterior to the tragus, supplying the concha, inferior crus, triangular fossa, antihelix and the earlobe." The stu
dy confirmed the constant presence of the helical rim arcade (Zilinsky-Cotofana), consistent perforating branches of the PAA, and the helical arcade (Erdman). |
https://en.wikipedia.org/wiki/Wolfgang%20M.%20Schmidt | Wolfgang M. Schmidt (born 3 October 1933) is an Austrian mathematician working in the area of number theory. He studied mathematics at the University of Vienna, where he received his PhD, which was supervised by Edmund Hlawka, in 1955. Wolfgang Schmidt is a Professor Emeritus from the University of Colorado at Boulder and a member of the Austrian Academy of Sciences and the Polish Academy of Sciences.
Career
He was awarded the eighth Frank Nelson Cole Prize in Number Theory for work on Diophantine approximation. He is known for his subspace theorem.
In 1960, he proved that every normal number in base r is normal in base s if and only if log r / log s is a rational number. He also proved the existence of T numbers. His series of papers on irregularities of distribution can be seen in J.Beck and W.Chen, Irregularities of Distribution, Cambridge University Press.
Schmidt is in a small group of number theorists who have been invited to address the International Congress of Mathematicians three times. The others are Iwaniec, Shimura, and Tate.
In 1986, Schmidt received the Humboldt Research Award and in 2003, he received the Austrian Decoration for Science and Art. Schmidt holds honorary doctorates from the University of Ulm, the Sorbonne, the University of Waterloo, the University of Marburg and the University of York. In 2012 he became a fellow of the American Mathematical Society.
Books
Diophantine approximation. Lecture Notes in Mathematics 785. Springer. (1980 [1996 with minor corrections])
Diophantine approximations and Diophantine equations, Lecture Notes in Mathematics, Springer Verlag 2000
Equations Over Finite Fields: An Elementary Approach, 2nd edition, Kendrick Press 2004 |
https://en.wikipedia.org/wiki/List%20of%20chatbots | This is a list of notable chatbots.
Chatbots are text-based conversation agents that can interact with human users through some medium, such as an instant message service. Some chatbots are designed for specific purposes, while others converse with human users on a wide range of topics.
Classic chatbots
Dr. Sbaitso
ELIZA
PARRY
Racter (or Claude Chatterbot)
Mark V Shaney
General chatbots
Albert One, 1998 and 1999 Loebner winner developed by Robby Garner
A.L.I.C.E., 2001, 2002, and 2004 Loebner Prize winner developed by Richard Wallace
Bard, AI chatbot developed by Google
Braina, intelligent personal assistant developed by Brainasoft
CarynAI, AI chatbot developed by Forever Voices
ChatGPT, prototype AI chatbot developed by OpenAI
Claude, chatbot developed by Anthropic
Cleverbot, winner of the 2010 Mechanical Intelligence Competition
Ernie Bot, chatbot developed by Baidu
Eugene Goostman, 2012 Turing 100 winner developed by Vladimir Veselov
Fred, early chatterbot developed by Robby Garner
Jabberwacky, chatterbot created by Rollo Carpenter
Jeeney AI, chatterbot from 2009
Lenny, audio bot designed to annoy telemarketers
MegaHAL, chatterbot developed by Jason Hutchens
SimSimi, a popular AI conversation program developed in 2002 by ISMaker
Sparrow, chatbot developed by DeepMind
Spookitalk, chatterbot used for NPCs in Douglas Adams' Starship Titanic video game
Tongyi Qianwen, chatbot developed by Alibaba Group
Ultra Hal, 2007 Loebner Prize winner developed by Robert Medeksza
Verbot, chatbot SDK for Windows
IM chatbots
GooglyMinotaur, specializing in Radiohead, the first bot released by ActiveBuddy (June 2001-March 2002)
SmarterChild, developed by ActiveBuddy and released in June 2001
Infobot, an assistant on IRC channels such as #perl, primarily to help out with answering Frequently Asked Questions (June 1995-today)
Chatbots in other languages
GigaChat, Russian chatbot developed by Sberbank
See also
The Pile (dataset), public data used to tra |
https://en.wikipedia.org/wiki/Automatic%20volume%20limiter%20system | An automatic volume limiter system (AVLS) is an option that limits the maximum volume level and is enabled through software or hardware in stationary or portable media player devices used with headphones such as the Walkman or Sony PSP. The aim of this feature is to stop the headphones drowning out all other noise, and to limit the noise from the headphones being heard by other people. It can also prevent listeners from damaging their hearing. This volume limit can be set by the manufacturer or customized by the user. |
https://en.wikipedia.org/wiki/Biocommunication%20%28science%29 | In the study of the biological sciences, biocommunication is any specific type of communication within (intraspecific) or between (interspecific) species of plants, animals, fungi, protozoa and microorganisms. Communication basically means sign-mediated interactions following three levels of (syntactic, pragmatic and semantic) rules. Signs in most cases are chemical molecules (semiochemicals), but also tactile, or as in animals also visual and auditive. Biocommunication of animals may include vocalizations (as between competing bird species), or pheromone production (as between various species of insects), chemical signals between plants and animals (as in tannin production used by vascular plants to warn away insects), and chemically mediated communication between plants and within plants.
Biocommunication of fungi demonstrates that mycelia communication integrates interspecific sign-mediated interactions between fungal organisms soil bacteria and plant root cells without which plant nutrition could not be organized. Biocommunication of Ciliates identifies the various levels and motifs of communication in these unicellular eukaryotes. Biocommunication of Archaea represents keylevels of sign-mediated interactions in the evolutionarily oldest akaryotes. Biocommunication of Phages demonstrates that the most abundant living agents on this planet coordinate and organize by sign-mediated interactions. Biocommunication is the essential tool to coordinate behavior of various cell types of immune systems.
Biocommunication, biosemiotics and linguistics
Biocommunication theory may be considered to be a branch of biosemiotics. Whereas Biosemiotics studies the production and interpretation of signs and codes, biocommunication theory investigates concrete interactions mediated by signs. Accordingly, syntactic, semantic, and pragmatic aspects of biocommunication processes are distinguished. Biocommunication specific to animals (animal communication) is considered a branch of |
https://en.wikipedia.org/wiki/Information%20Presentation%20Facility | Information Presentation Facility (IPF) is a system for presenting online help and hypertext on IBM OS/2 systems. IPF also refers to the markup language that is used to create IPF content. The IPF language has its origins in BookMaster and Generalized Markup Language developed by IBM. The IPF language is very similar to the well-known HTML language, version 3.0, with a range of additional possibilities. Therefore, a trained user may use virtually any word processor when creating IPF documents. The IPF language consists of 45 basic commands.
IPF files are compiled using the IPF Compiler (IPFC) into viewable INF or HLP files. IPF HLP files are distinct from the WinHelp HLP files that are prevalent in Windows.
OS/2 contains a built in viewer, and there are other viewers available for other platforms.
Example 1 - IBM
Here is a sample of IPF markup from IBM's Information Presentation Facility Programming Guide.
<nowiki>
.* This is a comment line
:userdoc.
:title.Endangered Mammals
:h1 res=001. The Manatee
:p.
The manatee has a broad flat tail and two flipper
like forelegs. There are no back legs.
The manatee's large upper lip is split in two and
can be used like fingers to place food into the
mouth. Bristly hair protrudes from its lips,
and almost buried in its hide are small eyes, with
which it can barely see.
:euserdoc.
</nowiki>
Example 2 - PM123 User's Manual
<nowiki>
:lm margin=2.:font facename=Helv size=24x10.
:p.:hp8.Welcome to PM123 !:ehp8.
:font facename=Helv size=16x8.
:p.:p.
Hello and welcome to the wonderful world of digital music on OS/2. First we
must congratulate you for choosing the best MPEG-audio player available for
OS/2! PM123 has been in development since beginning of 1997 and has become
the most advanced player on OS/2. Some of you may have used the earlier
betas of PM123 and for your convenience, here are the new features in this
release:
.br
:ul compact.
:li. New skin o |
https://en.wikipedia.org/wiki/Ultraviolet%20photoelectron%20spectroscopy | Ultraviolet photoelectron spectroscopy (UPS) refers to the measurement of kinetic energy spectra of photoelectrons emitted by molecules which have absorbed ultraviolet photons, in order to determine molecular orbital energies in the valence region.
Basic theory
If Albert Einstein's photoelectric law is applied to a free molecule, the kinetic energy () of an emitted photoelectron is given by
,
where h is Planck's constant, ν is the frequency of the ionizing light, and I is an ionization energy for the formation of a singly charged ion in either the ground state or an excited state. According to Koopmans' theorem, each such ionization energy may be identified with the energy of an occupied molecular orbital. The ground-state ion is formed by removal of an electron from the highest occupied molecular orbital, while excited ions are formed by removal of an electron from a lower occupied orbital.
History
Prior to 1960, virtually all measurements of photoelectron kinetic energies were for electrons emitted from metals and other solid surfaces. In about 1956, Kai Siegbahn developed X-ray photoelectron spectroscopy (XPS) for surface chemical analysis. This method uses x-ray sources to study energy levels of atomic core electrons, and at the time had an energy resolution of about 1 eV (electronvolt).
The ultraviolet photoelectron spectroscopy (UPS) was pioneered by Feodor I. Vilesov, a physicist at St. Petersburg (Leningrad) State University in Russia (USSR) in 1961 to study the photoelectron spectra of free molecules in the gas phase. The early experiments used monochromatized radiation from a hydrogen discharge and a retarding potential analyzer to measure the photoelectron energies.
The PES was further developed by David W. Turner, a physical chemist at Imperial College in London and then at Oxford University, in a series of publications from 1962 to 1967. As a photon source, he used a helium discharge lamp which emits a wavelength of 58.4 nm (corresponding to an ene |
https://en.wikipedia.org/wiki/Jasomi%20Networks | Jasomi Networks was a pioneer in the development of Session Border Controllers (SBCs), computer network devices that enable, control, and monitor the flow of multimedia data streams across carrier networks, corporate networks, home networks, and the Internet.
History and evolution
The concept for Jasomi was conceived in a meeting between Cullen Jennings and David A. Bryan, who introduced the concept to Dan Freedman, resulting in the founding of the corporation in 2001. Funded by the founders and early employee Benjamin Freedman, the company quickly created a demonstrable SBC, and began to market it as the PeerPoint product line at the Fall 2001 VON trade show organized by Jeff Pulver.
Strong early interest encouraged Freedman and Jennings to commit additional bootstrapping funds from themselves and friends and family. An R&D center was established in Calgary, Alberta, Canada in early 2002, and Alan Hawrylyshen joined the team. 2002 was spent turning the early demonstrator into a more full featured and supportable product. Ryan Kereliuk and Johnson Wu joined the company in mid-2002, and in late 2002, Jasomi adopted a hybrid coopetition model of software development, keeping certain advancements proprietary while releasing others to the public through the reSIProcate open-source SIP stack. David A. Bryan left the company and Alan Hawrylyshen became CTO in late 2002.
Between 2002 and 2004, the company advanced the state of the art in SBC technology, providing customers for the first time with the ability to perform VoIP through existing NAT-enabled firewalls so that residential subscribers could be supported without placing any hardware on their premises. An early customer making great use of this facility was Jeff Pulver's Free World Dialup, which provided free calling services worldwide amongst its subscribers using the SIP VoIP protocol.
In 2005, the company was acquired by Ditech Communications (itself later acquired by Nuance Communications) in a deal valued |
https://en.wikipedia.org/wiki/Video%20interlude | A video interlude is an interlude during a performance that shows a video. Video interludes are often played in concerts, showing a music video (often made specifically for the show), usually featuring the artists while the artist takes a break or costume change.
Video interludes have been used by Madonna since at least 1990. |
https://en.wikipedia.org/wiki/Sky%20and%20Water%20I | Sky and Water I is a woodcut print by the Dutch artist M. C. Escher first printed in June 1938. The basis of this print is a regular division of the plane consisting of birds and fish. Both prints have the horizontal series of these elements—fitting into each other like the pieces of a jigsaw puzzle—in the middle, transitional portion of the prints. In this central layer the pictorial elements are equal: birds and fish are alternately foreground or background, depending on whether the eye concentrates on light or dark elements. The birds take on an increasing three-dimensionality in the upward direction, and the fish, in the downward direction. But as the fish progress upward and the birds downward they gradually lose their shapes to become a uniform background of sky and water, respectively.
According to Escher: "In the horizontal center strip there are birds and fish equivalent to each other. We associate flying with sky, and so for each of the black birds the sky in which it is flying is formed by the four white fish which encircle it. Similarly swimming makes us think of water, and therefore the four black birds that surround a fish become the water in which it swims."
This print has been used in physics, geology, chemistry, and in psychology for the study of visual perception. In the pictures a number of visual elements unite into a simple visual representation, but separately each forms a point of departure for the elucidation of a theory in one of these disciplines.
See also
Sky and Water II
Tessellation
Sources
M. C. Escher—The Graphic Work; Benedikt-Taschen Publishers.
M. C. Escher—29 Master Prints; Harry N. Abrams, Inc., Publishers.
Locher, J. L. (2000). The Magic of M. C. Escher. Harry N. Abrams, Inc. .
Works by M. C. Escher
1938 works
Woodcuts
Fish in art
Birds in art
he:שמים ומים#שמים ומים 1 |
https://en.wikipedia.org/wiki/Rewritable%20consumer%20timecode | The Rewriteable Consumer Timecode (RCTC, RC Timecode, or RC Time Code) is a nearly frame accurate timecode method developed by Sony for 8mm and Hi8 analog tape formats. The RC timecode tags each frame with the hour, minute, second and frame for each frame of video recorded to tape. Officially, RCTC is accurate to within ±2 to 5 frames. The RC timecode can be used in conjunction with the datacode to record the date and the time. The data and RC codes are written between the video and the PCM audio tracks. It may be added to any 8-mm tape without altering the information already on the tape, and is invisible to machines not equipped to read it.
On several video camera models, Sony included the ability to search by date/time (appropriately called Date Search) and to index positions within the tape so that the index mark could be returned to with the press of a button.
The RC timecode is a different technology than the SMPTE timecode, linear timecode (LTC), and vertical interval timecode (VITC).
At least one manufacturer (Octochron) created tools to capture the information from the RC timecode on tape and export it via Serial or USB to the computer.
A partial list of camcorders which supported RC timecode.
See also
Linear timecode (LTC)
Vertical interval timecode (VITC)
CTL timecode
MIDI timecode
AES-EBU embedded timecode |
https://en.wikipedia.org/wiki/Carotid%20sinus%20nerve | The carotid branch of the glossopharyngeal nerve (carotid sinus nerve or Hering's nerve) is a small branch of the glossopharyngeal nerve (cranial nerve IX) that innervates the carotid sinus, and carotid body.
Anatomy
Course and relations
It runs downward anterior to the internal carotid artery. It communicates with the vagus nerve and sympathetic trunk before dividing in the angle of the bifurcation of the common carotid artery to innervate the carotid body, and carotid sinus.
Function
It conveys information from the baroreceptors of the carotid sinus to the vasomotor center in the brainstem (in order to mediate blood pressure homeostasis), and from chemoreceptors of the carotid body (mainly conveying information about partial pressures of blood oxygen, and carbon dioxide). |
https://en.wikipedia.org/wiki/Gold%E2%80%93aluminium%20intermetallic | A gold–aluminium intermetallic is an intermetallic compound of gold and aluminium that occurs at contacts between the two metals.
These intermetallics have different properties from the individual metals, which can cause problems in wire bonding in microelectronics. The main compounds formed are Au5Al2 (white plague) and AuAl2 (purple plague), which both form at high temperatures.
White plague is the name of the compound Au5Al2 as well as the problem it causes. It has low electrical conductivity, so its formation at the joint leads to an increase of electrical resistance which can lead to total failure. Purple plague (sometimes known as purple death or Roberts-Austen's purple gold) is a brittle, bright-purple compound, AuAl2, or about 78.5% Au and 21.5% Al by mass. AuAl2 is the most stable thermally of the Au–Al intermetallic compounds, with a melting point of 1060°C (see phase diagram), similar to that of pure gold. The process of the growth of the intermetallic layers causes reduction in volume, and hence creates cavities in the metal near the interface between gold and aluminium.
Other gold–aluminium intermetallics can cause problems as well. Below 624°C, purple plague is replaced by Au2Al, a tan-colored substance. It is a poor conductor and can cause electrical failure of the joint that can lead to mechanical failure. At lower temperatures, about 400–450°C, an interdiffusion process takes place at the junction. This leads to formation of layers of several intermetallic compounds with different compositions, from gold-rich to aluminium-rich, with different growth rates. Cavities form as the denser, faster-growing layers consume the slower-growing ones. This process, known as Kirkendall voiding, leads to both increased electrical resistance and mechanical weakening of the wire bond. When the voids are collected along the diffusion front, a process aided by contaminants present in the lattice, it is known as Horsting voiding, a process similar to and often con |
https://en.wikipedia.org/wiki/List%20of%20pop-up%20blocking%20software | This is a list of software that can block pop-up ads. Blocking is usually a user-enabled option, and can in many cases allow specified exceptions.
Browsers that can block pop-up ads
Trident shells
AOL Explorer
Avant Browser
GreenBrowser
Internet Explorer
Lunascape
Maxthon
MSN Explorer
NeoPlanet
Netcaptor
Netscape 8
Sleipnir
Gecko-based browsers
Camino
Epiphany
Flock
Galeon
K-Meleon
Lunascape
Mozilla Application Suite
Mozilla Firefox
Netscape 7
Netscape 8
SeaMonkey
KHTML/WebKit-based browsers
Brave
Google Chrome
iCab
Konqueror
Lunascape
OmniWeb
Safari
Shiira
Presto-based browsers
Opera
Others
Links
NetSurf
w3m
Add-on programs that block pop-up ads
Privoxy
Proxomitron
Browser extensions
Adblock Plus
AdBlock
Adguard
Alexa Toolbar
Bing Bar
Google Toolbar
Ghostery
IE7pro
iMacros
NoScript—open source (GPL)
uBlock Origin
Yahoo! Toolbar |
https://en.wikipedia.org/wiki/List%20of%20books%20in%20computational%20geometry | This is a list of books in computational geometry.
There are two major, largely nonoverlapping categories:
Combinatorial computational geometry, which deals with collections of discrete objects or defined in discrete terms: points, lines, polygons, polytopes, etc., and algorithms of discrete/combinatorial character are used
Numerical computational geometry, also known as geometric modeling and computer-aided geometric design (CAGD), which deals with modelling of shapes of real-life objects in terms of curves and surfaces with algebraic representation.
Combinatorial computational geometry
General-purpose textbooks
The book is the first comprehensive monograph on the level of a graduate textbook to systematically cover the fundamental aspects of the emerging discipline of computational geometry. It is written by founders of the field and the first edition covered all major developments in the preceding 10 years. In the aspect of comprehensiveness it was preceded only by the 1984 survey paper, Lee, D, T., Preparata, F. P. : "Computational geometry - a survey". IEEE Trans. on Computers. Vol. 33, No. 12, pp. 1072–1101 (1984). It is focused on two-dimensional problems, but also has digressions into higher dimensions.
The initial core of the book was M.I.Shamos' doctoral dissertation, which was suggested to turn into a book by a yet another pioneer in the field, Ronald Graham.
The introduction covers the history of the field, basic data structures, and necessary notions from the theory of computation and geometry.
The subsequent sections cover geometric searching (point location, range searching), convex hull computation, proximity-related problems (closest points, computation and applications of the Voronoi diagram, Euclidean minimum spanning tree, triangulations, etc.), geometric intersection problems, algorithms for sets of isothetic rectangles
The monograph is a rather advanced exposition of problems and approaches in computational geometry focused on the role of |
https://en.wikipedia.org/wiki/Global%20Health%20Security%20Initiative | The Global Health Security Initiative (GHSI) is an informal international partnership among countries in order to exchange information and coordinate practices for confronting new threats and risks to global health. It was formed to respond to threats of biological, chemical, or radio-nuclear terrorism (CBRN), with pandemic influenza added to the scope a year later.
History
The idea on which the Global Health Security Initiative is based was suggested by then US Secretary of Health and Human Services, Tommy Thompson, after the World Trade Center attacks on 11 September 2001. He proposed that countries fighting bioterrorism should collaborate, share information and coordinate their efforts in order to best protect global health.
GHSI was launched in November 2001 by Canada (who hosted the first meeting in Ottawa), the European Commission, France, Germany, Italy, Japan, Mexico, the United Kingdom and the United States. The World Health Organization (WHO) would act as observer to the GHSI. The ministers agreed on eight areas in which the partnership could collaborate in order to "strengthen public health preparedness and response to the threat of international biological, chemical and radio-nuclear terrorism".
In December 2002 at a meeting in Mexico City, the Ministers broadened the scope of the mandate to include the public health threat posed by pandemic influenza.
Aims and scope
GHSI states that its mandate is "to undertake concerted global action to strengthen public health preparedness and response to chemical, biological, radiological, and nuclear (CBRN) threats, as well as pandemic influenza", which include intentional, accidental and naturally occurring events.
Organization
The Global Health Security Action Group (GHSAG) is made up of senior officials from each member country. The GHSI Secretariat organises, manages and administers meetings and committees and sets priorities.
Various technical/scientific working groups focus on specific areas of knowledg |
https://en.wikipedia.org/wiki/Zero-forcing%20equalizer | The zero-forcing equalizer is a form of linear equalization algorithm used in communication systems which applies the inverse of the frequency response of the channel. This form of equalizer was first proposed by Robert Lucky.
The zero-forcing equalizer applies the inverse of the channel frequency response to the received signal, to restore the signal after the channel. It has many useful applications. For example, it is studied heavily for IEEE 802.11n (MIMO) where knowing the channel allows recovery of the two or more streams which will be received on top of each other on each antenna. The name zero-forcing corresponds to bringing down the intersymbol interference (ISI) to zero in a noise-free case. This will be useful when ISI is significant compared to noise.
For a channel with frequency response the zero-forcing equalizer is constructed by . Thus the combination of channel and equalizer gives a flat frequency response and linear phase .
In reality, zero-forcing equalization does not work in most applications, for the following reasons:
Even though the channel impulse response has finite length, the impulse response of the equalizer needs to be infinitely long
At some frequencies the received signal may be weak. To compensate, the magnitude of the zero-forcing filter ("gain") grows very large. As a consequence, any noise added after the channel gets boosted by a large factor and destroys the overall signal-to-noise ratio. Furthermore, the channel may have zeros in its frequency response that cannot be inverted at all. (Gain * 0 still equals 0).
This second item is often the more limiting condition. These problems are addressed in the linear MMSE equalizer by making a small modification to the denominator of : , where k is related to the channel response and the signal SNR.
Algorithm
If the channel response (or channel transfer function) for a particular channel is H(s) then the input signal is multiplied by the reciprocal of it. This is intended to r |
https://en.wikipedia.org/wiki/Outline%20of%20life%20extension | The following outline is provided as an overview of and topical guide to life extension:
Life extension – study of slowing down or reversing the processes of aging to extend both the maximum and average lifespan. Also known as anti-aging medicine, experimental gerontology, and biomedical gerontology.
Goals
Biological immortality
Indefinite lifespan
Longevity
Rejuvenation
Longevity
Centenarian
Supercentenarian
Longevity escape velocity
Strategies
Research and development
Timeline of senescence research
Available strategies
Calorie restriction or protein restriction or intermittent fasting
Exercise
Geroprotector
Senolytics
Potential future strategies
Cryonics
Genetic therapies
Cloning and body part replacement
Cell replacement therapies (CRT)
Strategies for engineered negligible senescence (SENS)
Digital immortality through mind uploading
Suspended animation
Immunisation
Aging
Evolution of ageing
Gerontology
Pro-aging trance
Hallmarks of aging
Altered intercellular communication
Cellular senescence
Deregulated nutrient sensing
Epigenetic alterations
Genomic instability
Loss of proteostasis
Mitochondrial dysfunction
Stem cell exhaustion
Telomere attrition
Causes of aging
Cross-links
Crosslinking of DNA
Free radicals – atom, molecule, or ion that has unpaired valence electrons. With some exceptions, these unpaired electrons make free radicals highly chemically reactive towards other substances, or even towards themselves: their molecules will often spontaneously dimerize or polymerize if they come in contact with each other. They are countered to some extent by antioxidants.
Glycation
Advanced glycation end-product
Lipofuscin
Viral infections (acute or chronic)
Theories of aging
Theories of aging
Antagonistic pleiotropy theory of aging
Caloric restriction theory
Cross-linkage theory of aging
Death hormone theory
Disposable soma theory of aging
DNA damage theory of aging
Epigenetic clock theory of a |
https://en.wikipedia.org/wiki/Indra%27s%20Pearls%20%28book%29 | Indra's Pearls: The Vision of Felix Klein is a geometry book written by David Mumford, Caroline Series and David Wright, and published by Cambridge University Press in 2002 and 2015.
The book explores the patterns created by iterating conformal maps of the complex plane called Möbius transformations, and their connections with symmetry and self-similarity. These patterns were glimpsed by German mathematician Felix Klein, but modern computer graphics allows them to be fully visualised and explored in detail.
Title
The book's title refers to Indra's net, a metaphorical object described in the Buddhist text of the Flower Garland Sutra. Indra's net consists of an infinite array of gossamer strands and pearls. The frontispiece to Indra's Pearls quotes the following description:
In the glistening surface of each pearl are reflected all the other pearls ... In each reflection, again are reflected all the infinitely many other pearls, so that by this process, reflections of reflections continue without end.
The allusion to Felix Klein's "vision" is a reference to Klein's early investigations of Schottky groups and hand-drawn plots of their limit sets. It also refers to Klein's wider vision of the connections between group theory, symmetry and geometry - see Erlangen program.
Contents
The contents of Indra's Pearls are as follows:
Chapter 1. The language of symmetry – an introduction to the mathematical concept of symmetry and its relation to geometric groups.
Chapter 2. A delightful fiction – an introduction to complex numbers and mappings of the complex plane and the Riemann sphere.
Chapter 3. Double spirals and Möbius maps – Möbius transformations and their classification.
Chapter 4. The Schottky dance – pairs of Möbius maps which generate Schottky groups; plotting their limit sets using breadth-first searches.
Chapter 5. Fractal dust and infinite words – Schottky limit sets regarded as fractals; computer generation of these fractals using depth-first searches and i |
https://en.wikipedia.org/wiki/VB%20Watch | VB Watch is a Visual Basic programming utility. VB Watch consists of three tools for Visual Basic 6.0: Profiler, Protector and Debugger.
Profiler
VB Watch Profiler measures the speed of a running Visual Basic program. It displays the time spent in each procedure and/or a line of code. This information can be used in code optimization to detect bottleneck procedures and lines. The Profiler can also be used to measure code coverage during software testing.
Debugger
VB Watch Debugger monitors what happens inside a running Visual Basic program or library. It displays the call stack, execution trace, global variables and the number of live objects.
The Debugger also allows one to add breakpoints in executable files.
External links
VB Watch home page
Debuggers
Profilers
Software testing tools
Programming tools for Windows |
https://en.wikipedia.org/wiki/Control%20message | Control messages are a special kind of Usenet post that are used to control news servers. They differ from ordinary posts by a header field named Control. The body of the field contains control name and arguments.
There are two historical alternatives to header field Control. They are not supported by contemporary software and forbidden according to RFC 5537. However, the traditional format of the subject line is widely used in addition to the Control header: the subject line consists of the word "cmsg" followed by control name and arguments.
Types
cancel
A cancel message requests the deletion of a specific article. The body of the Control field contains one argument, the Message-ID of the article to delete.
According to RFC 1036 only the author of the target message or the local news administrator is allowed to send a cancel (cancels not meeting this condition are called "rogue cancels"). To verify authorization the line (or line, if it exists) of the cancel message must match the target article. This verification does not work well in modern day Usenet and is rarely used.
Additional hierarchy specific rules (see Breidbart Index) allow cancelbots to send third-party cancels to remove spam.
Example
Control: cancel <899qh19zehlhsdfa@example.com>
Subject: cmsg cancel <899qh19zehlhsdfa@example.com>
newgroup
A newgroup message is issued to create a new Usenet newsgroup. The body of the Control field contains one mandatory argument, the name of the new group. The second argument is optional. If present it consists of the keyword . The body of the message typically contains tagline, charter and rationale.
If the group already exists, only the status of the group is changed, i.e. whether it is moderated or nor not.
Typically newgroup messages having a correct digital signature are executed automatically. In some hierarchies (, , ) unsigned newgroup messages just serve as formalized proposal to create a new group. Objections to the proposal are then expressed |
https://en.wikipedia.org/wiki/Chattering%20teeth | Chattering teeth is a bodily function in animals that occurs primarily in response to cold; the jaw muscles begin to shiver leading teeth to crash together. It may also occur as a result of bruxism where emotional stress causes the jaw movements. Certain medications can lead to teeth chattering as a side effect, especially antipsychotics and selective serotonin reuptake inhibitors. |
https://en.wikipedia.org/wiki/Towers%20Perrin | Towers Perrin was a professional services firm specializing in human resources and financial services consulting, which merged in January 2010 with Watson Wyatt to form Towers Watson. The firm was a provider of reinsurance intermediary services and was active in the actuarial consulting arena with its Tillinghast insurance consulting practice.
History
On March 1, 1934, Towers, Perrin, Forster & Crosby opened for business in Philadelphia, Pennsylvania. In 1987, Towers Perrin was established as the umbrella name for the firm. On June 28, 2009, Towers, Perrin, Forster & Crosby and Watson Wyatt Worldwide (NYSE:WW) announced their agreement to merge into a new publicly traded company called Towers Watson & Co. The merger was completed in January 2010.
Offices
Towers Perrin had offices and alliance partners in the United States, Canada, Europe, Asia, Latin America, South Africa, Australia and New Zealand. |
https://en.wikipedia.org/wiki/Statistical%20shape%20analysis | Statistical shape analysis is an analysis of the geometrical properties of some given set of shapes by statistical methods. For instance, it could be used to quantify differences between male and female gorilla skull shapes, normal and pathological bone shapes, leaf outlines with and without herbivory by insects, etc. Important aspects of shape analysis are to obtain a measure of distance between shapes, to estimate mean shapes from (possibly random) samples, to estimate shape variability within samples, to perform clustering and to test for differences between shapes. One of the main methods used is principal component analysis (PCA). Statistical shape analysis has applications in various fields, including medical imaging, computer vision, computational anatomy, sensor measurement, and geographical profiling.
Landmark-based techniques
In the point distribution model, a shape is determined by a finite set of coordinate points, known as landmark points. These landmark points often correspond to important identifiable features such as the corners of the eyes. Once the points are collected some form of registration is undertaken. This can be a baseline methods used by Fred Bookstein for geometric morphometrics in anthropology. Or an approach like Procrustes analysis which finds an average shape.
David George Kendall investigated the statistical distribution of the shape of triangles, and represented each triangle by a point on a sphere. He used this distribution on the sphere to investigate ley lines and whether three stones were more likely to be co-linear than might be expected. Statistical distribution like the Kent distribution can be used to analyse the distribution of such spaces.
Alternatively, shapes can be represented by curves or surfaces representing their contours, by the spatial region they occupy.
Shape deformations
Differences between shapes can be quantified by investigating deformations transforming one shape into another. In particular a diffe |
https://en.wikipedia.org/wiki/Robert%20Rayford | Robert Lee Rayford (February 3, 1953 – May 15, 1969), sometimes identified as Robert R. due to his age, was an American teenager from Missouri who has been suggested to represent the earliest confirmed case of HIV/AIDS in North America. This is based on evidence published in 1988 in which the authors claimed that medical evidence indicated that he was "infected with a virus closely related or identical to human immunodeficiency virus type 1." Rayford died of pneumonia, but his other symptoms baffled the doctors who treated him. A study published in 1988 reported the detection of antibodies against HIV. Results of testing for HIV genetic material were reported once at a scientific conference in Australia in 1999; however, the data has never been published in a peer-reviewed medical or scientific journal.
Background
Robert Rayford was born on February 3, 1953, in St. Louis, Missouri. As a single parent, Constance had to raise both children by herself. The Rayfords lived in the Old North neighborhood of St. Louis, where the 19th-century brick homes provided affordable housing for several working class African-American families such as their own. Not much of Rayford's personal life was disclosed.
Illness
In early 1968, Rayford, then 15 years old, admitted himself to the City Hospital in St. Louis. His legs and genitals were covered in warts and sores. He also had severe swelling of the testicles and pelvic region, which later spread to his legs, causing a misdiagnosis of lymphedema.
He had grown thin and pale and suffered from shortness of breath. Rayford told the doctors that he had experienced these symptoms since at least late 1966.
Tests discovered a severe chlamydia infection which had, unusually, spread throughout his body. Rayford declined a rectal examination request from hospital personnel, and was described as uncommunicative and withdrawn.
Dr. Memory Elvin-Lewis, who was assigned to his case, would recall of his shy and somewhat hesitant personalit |
https://en.wikipedia.org/wiki/Asiatic%20Lion%20Reintroduction%20Project | The Asiatic Lion Reintroduction Project is an initiative of the Indian Government to provide safeguards to the Asiatic lion (Panthera leo leo) from extinction in the wild by means of reintroduction. The last wild population of the Asiatic lion is found in the region of Gir Forest National Park, in the state of Gujarat. The single population faces the threats of epidemics, natural disasters and other anthropogenic factors. The project aims to establish a second independent population of Asiatic lions at the Kuno National Park in the Indian state of Madhya Pradesh. However, the proposed translocation has been bitterly contested by the state government.
History
The distribution of Asiatic lion, once found widely in West and South Asia, dwindled to a single population in the Gir Forest National Park and Wildlife Sanctuary in India. The population at Gir declined to 18 animals in 1893 but increased due to protection and conservation efforts to 284 in 1994. The Gir Wildlife Sanctuary is now highly overpopulated with lions, the 2015 census showed the strength to be 523 lions. There are numerous deaths in the population annually because of ever increasing competition between the human and animal overcrowding. Asiatic lion prides require large territories but there is limited space at Gir wildlife sanctuary, which is boxed in on all sides by heavy human habitation. Gir lions have started moving outwards from the sanctuary and establishing homes outside the protected areas. The lions are now spread over in the vicinity of 1050 villages in four contiguous districts - Amreli, Gir Somnath , Bhavnagar and Junagadh.
Sheopur introduction of 1904
The Maharaja of Gwalior, on being encouraged by Lord Curzon in 1904, imported cubs of African lions and attempted to introduce them in the wild in the forests near Sheopur. The introduced lions took to raiding livestock and some even turned to man-eating, subsequent to which they were all eventually tracked down and shot.
The Chandrap |
https://en.wikipedia.org/wiki/Electronic%20circuit | An electronic circuit is composed of individual electronic components, such as resistors, transistors, capacitors, inductors and diodes, connected by conductive wires or traces through which electric current can flow. It is a type of electrical circuit and to be referred to as electronic, rather than electrical, generally at least one active component must be present. The combination of components and wires allows various simple and complex operations to be performed: signals can be amplified, computations can be performed, and data can be moved from one place to another.
Circuits can be constructed of discrete components connected by individual pieces of wire, but today it is much more common to create interconnections by photolithographic techniques on a laminated substrate (a printed circuit board or PCB) and solder the components to these interconnections to create a finished circuit. In an integrated circuit or IC, the components and interconnections are formed on the same substrate, typically a semiconductor such as doped silicon or (less commonly) gallium arsenide.
An electronic circuit can usually be categorized as an analog circuit, a digital circuit, or a mixed-signal circuit (a combination of analog circuits and digital circuits). The most widely used semiconductor device in electronic circuits is the MOSFET (metal–oxide–semiconductor field-effect transistor).
Breadboards, perfboards, and stripboards are common for testing new designs. They allow the designer to make quick changes to the circuit during development.
Analog circuits
Analog electronic circuits are those in which current or voltage may vary continuously with time to correspond to the information being represented.
The basic components of analog circuits are wires, resistors, capacitors, inductors, diodes, and transistors. Analog circuits are very commonly represented in schematic diagrams, in which wires are shown as lines, and each component has a unique symbol. Analog circuit analys |
https://en.wikipedia.org/wiki/List%20of%20biogeographic%20provinces | This page features a list of biogeographic provinces that were developed by Miklos Udvardy in 1975, later modified by other authors. Biogeographic Province is a biotic subdivision of biogeographic realms subdivided into ecoregions, which are classified based on their biomes or habitat types and, on this page, correspond to the floristic kingdoms of botany.
The provinces represent the large areas of Earth's surface within which organisms have been evolving in relative isolation over long periods of time, separated from one another by geographic features, such as oceans, broad deserts, or high mountain ranges, that constitute barriers to migration.
Biomes are characterized by similar climax vegetation, though each realm may include a number of different biomes. A tropical moist broadleaf forest in Brazil, for example, may be similar to one in New Guinea in its vegetation type and structure, climate, soils, etc., but these forests are inhabited by plants with very different evolutionary histories.
Afrotropical Realm
Tropical humid forests
Guinean Rainforest
Congo Rainforest
Malagasy Rainforest
Tropical dry or deciduous forests (incl. Monsoon forests) or woodlands
West African Woodland/Savanna
East African Woodland/Savanna
Congo Woodland/Savanna
Miombo Woodland/Savanna
South African Woodland/Savanna
Malagasy Woodland/Savanna
Malagasy Thorn Forest
Evergreen sclerophyllous forests, scrubs or woodlands
Cape Sclerophyll
Warm deserts and semideserts
Western Sahel
Eastern Sahel
Somalian
Namib
Kalahari
Karroo
Mixed mountain and highland systems with complex zonation
Ethiopian Highlands
Guinean Highlands
Central African Highlands
East African Highlands
South African Highlands
Mixed island systems
Ascension and St. Helena Islands
Comores Islands and Aldabra
Mascarene Islands
Lake systems
Lake Rudolph
Lake Ukerewe (Victoria)
Lake Tanganyika
Lake Malawi (Nyassa)
Antarctic Realm
Tundra communities and barren Antarctic desert
Subtropical a |
https://en.wikipedia.org/wiki/Anatomical%20neck%20of%20humerus | The anatomical neck of the humerus is obliquely directed, forming an obtuse angle with the body of the humerus. It represents the fused epiphyseal plate.
Structure
The anatomical neck divides the head of the humerus from the greater and lesser tubercles of the humerus. It gives attachment to the capsular ligament of the shoulder joint except at the upper inferior-medial aspects. It is best marked in the lower half of its circumference; in the upper half it is represented by a narrow groove separating the head of the humerus from the two tubercles, the greater tubercle and the lesser tubercle. It affords attachment to the articular capsule of the shoulder-joint, and is perforated by numerous vascular foramina.
Additional images |
https://en.wikipedia.org/wiki/Surgical%20neck%20of%20the%20humerus | The surgical neck of the humerus is a bony constriction at the proximal end of shaft of humerus. It is situated distal to the greater tubercle and lesser tubercle, and proximal to the deltoid tuberosity.
Clinical significance
The surgical neck is much more frequently fractured than the anatomical neck of the humerus. This type of fracture takes place when the humerus is forced in one direction while the joint capsule and the rotator cuff muscles remain intact. A fracture in this area is most likely to cause damage to the axillary nerve and posterior circumflex humeral artery. Damage to the axillary nerve affects function of the teres minor and deltoid muscles, resulting in loss of abduction of arm (from 15-90 degrees), weak flexion, extension, and rotation of shoulder as well as loss of sensation of the skin over a small part of the lateral shoulder.
Additional images
See also
Quadrangular space
Anatomical neck of the humerus |
https://en.wikipedia.org/wiki/GBDE | GBDE, standing for GEOM Based Disk Encryption, is a block device-layer disk encryption system written for FreeBSD, initially introduced in version 5.0. It is based on the GEOM disk framework. GBDE was designed and implemented by Poul-Henning Kamp and Network Associates Inc. (now known as McAfee).
Design decisions
Unlike most disk encryption software, GBDE does not attempt to defeat watermarking attacks through the use of disk encryption-specific modes of operation (see disk encryption theory), but instead generates a random key each time a sector is written. Unlike some alternatives, such as CBC with sector-specific initialization vectors, this approach does not reveal any information to the attacker even if they have access to snapshots of the disk image from different points in time, since encryption keys are never re-used.
The one time sector key is encrypted using a pseudorandom key. This pseudorandom key is derived from the sector number and a static 2048-bit master key with 128 bits of salt. The pseudorandom number generator used for this purpose is called the Cherry Picker. This is not a well established PRNG, but rather one invented for GBDE. This generator may not meet the security levels of standard algorithms, and could be distinguishable from random numbers.
Limitations
Due to this unique approach, GBDE only supports 128-bit AES. Using a different key for each write also introduces a significant CPU overhead, as most block ciphers use key-specific precomputations, and makes disk updates non-atomic since the keys are written separately from the data. As a result, data loss can occur on unexpected power drops, even when used with journaling file systems. GBDE also has a disk space overhead of about 3% to store the per-sector keys.
To address these shortcomings, a more typical disk encryption solution for FreeBSD, GELI, was written later by Pawel Jakub Dawidek.
See also
GELI
Disk encryption
Disk encryption software
Comparison of disk encryption |
https://en.wikipedia.org/wiki/Daniel%20McCartney | Daniel McCartney (September 10, 1817 – November 15, 1887) was an American who had (what is now known as) Highly Superior Autobiographical Memory (HSAM).
McCartney was born in Westmoreland County, Pennsylvania. He was legally blind and lived with relatives throughout his lifetime. For a large part of his life, he resided in Morrow County, Ohio, before his final days in Muscatine, Iowa. He never married.
McCartney was famous for his mental ability in two specific areas, memory and calculation.
He remembered every single day in his life from the age of nine until his death. Given any specific calendar date, McCartney could, in seconds, give the day of the week, describe the weather conditions, describe what he did during the day, describe what he ate during the day, and provide details of local, regional and national events on that day. In a similar case to McCartney's, noted neurobiologist James McGaugh of the University of California, Irvine, one of the world's leading experts on human memory, reported a woman, Jill Price, with the astonishing ability to clearly remember events that happened to her decades before. McGaugh labeled this one-of-a-kind ability as Hyperthymesia (National Public Radio, 2006). McCartney's mental aptitude appears to be nearly identical to this recent case reported in McGaugh's study. McCartney, however, had an additional mental ability: mathematical computation.
McCartney could mentally compute difficult mathematical computations in seconds, and extremely difficult ones in minutes. McCartney was tested several times by panels of university mathematicians in which he was given a battery of mathematical questions. On one such examination in July 1870 in Salem, Ohio, McCartney was asked to take 89 to the sixth power, which he mentally computed in ten minutes, giving the correct answer of 496,981,290,961. On another examination he was asked to provide the cube root of 4,741,632 for which he answered correctly in three minutes-(168); and 389, |
https://en.wikipedia.org/wiki/Radial%20fossa | The radial fossa is a slight depression found on the humerus above the front part of the capitulum. It receives the anterior border of the head of the radius when the forearm is flexed.
Structure
The joint capsule of the elbow attaches to the humerus just proximal to the radial fossa.
Additional images |
https://en.wikipedia.org/wiki/Telecom%20Gold | Telecom Gold (sometimes also known as BT Gold) was an early commercial electronic mail service launched by British Telecom in 1982. It was based on Prime minicomputers running Dialcom software under a customised version of PRIMOS. (ITT Dialcom was later acquired by BT in 1986.) The system offered various services, including e-mail to and from other Telecom Gold users and those of Dialcom services in other countries, and other e-mail systems such as Sprint and integration with telex, fax, online databases and an experimental OCR system for a short while. Later, X.400 functionality was added.
Users would dial into the system using a conventional modem and terminal emulator. Alternatively, users could dial a local number and connect via the PSS X.25 network. The X.400 services also had a Mail User Agent which ran on IBM PCs and compatibles.
The UK data centre was originally located in the basement of Beckett House 60-68 St Thomas St, Bermondsey, London, SE1 3QU but later moved to a custom built facility at Oxgate Centre, Oxgate Ln, London NW2 7JA which now houses LDEX1.
The service eventually became obsolete with the growth of the Internet in 1996.
Although BT continued to market the service, it decided not to develop its [Telecom Gold] successor, Mailbox, into an Internet Service Provider when it became clear that people wanted to connect to the Internet during the early to mid 1990s. Instead, BT decided to launch a new Internet Service Provider, called BTnet, in 1994, and within two years, Mailbox had ceased to exist.
During the 1980s, BT Gold hosted one of the first online communities. Users communicated using a noticeboard () and via a simple chat facility which allowed real-time conversations to take place. The BT Gold community was worldwide, but the majority of users were in London and would meet regularly at "eyeballs" (coined from CB usage).
See also
Robert Schifreen & Steve Gold, alleged hackers of Prince Philip's Telecom Gold mailbox in 1985 |
https://en.wikipedia.org/wiki/Transverse%20tarsal%20joint | The transverse tarsal joint or midtarsal joint or Chopart's joint is formed by the articulation of the calcaneus with the cuboid (the calcaneocuboid joint), and the articulation of the talus with the navicular (the talocalcaneonavicular joint).
The movement which takes place in this joint is more extensive than that in the other tarsal joints, and consists of a sort of rotation by means of which the foot may be slightly flexed or extended, the sole being at the same time carried medially (inverted) or laterally (everted).
The term Chopart's joint is named after the French surgeon François Chopart. |
https://en.wikipedia.org/wiki/Ulam%20number | In mathematics, the Ulam numbers comprise an integer sequence devised by and named after Stanislaw Ulam, who introduced it in 1964. The standard Ulam sequence (the (1, 2)-Ulam sequence) starts with U1 = 1 and U2 = 2. Then for n > 2, Un is defined to be the smallest integer that is the sum of two distinct earlier terms in exactly one way and larger than all earlier terms.
Examples
As a consequence of the definition, 3 is an Ulam number (1 + 2); and 4 is an Ulam number (1 + 3). (Here 2 + 2 is not a second representation of 4, because the previous terms must be distinct.) The integer 5 is not an Ulam number, because 5 = 1 + 4 = 2 + 3. The first few terms are
1, 2, 3, 4, 6, 8, 11, 13, 16, 18, 26, 28, 36, 38, 47, 48, 53, 57, 62, 69, 72, 77, 82, 87, 97, 99, 102, 106, 114, 126, 131, 138, 145, 148, 155, 175, 177, 180, 182, 189, 197, 206, 209, 219, 221, 236, 238, 241, 243, 253, 258, 260, 273, 282, ... .
There are infinitely many Ulam numbers. For, after the first n numbers in the sequence have already been determined, it is always possible to extend the sequence by one more element: is uniquely represented as a sum of two of the first n numbers, and there may be other smaller numbers that are also uniquely represented in this way, so the next element can be chosen as the smallest of these uniquely representable numbers.
Ulam is said to have conjectured that the numbers have zero density, but they seem to have a density of approximately 0.07398.
Properties
Apart from 1 + 2 = 3 any subsequent Ulam number cannot be the sum of its two prior consecutive Ulam numbers.
Proof: Assume that for n > 2, Un−1 + Un = Un+1 is the required sum in only one way; then so does Un−2 + Un produce a sum in only one way, and it falls between Un and Un+1. This contradicts the condition that Un+1 is the next smallest Ulam number.
For n > 2, any three consecutive Ulam numbers (Un−1, Un, Un+1) as integer sides will form a triangle.
Proof: The previous property states that for n > 2, Un−2 + |
https://en.wikipedia.org/wiki/Switched%20video | Switched video or switched digital video (SDV), sometimes referred to as switched broadcast (SWB), is a telecommunications industry term for a network scheme for distributing digital video via a cable. Switched video sends the digital video more efficiently freeing bandwidth. The scheme applies to digital video distribution both on typical cable TV systems using QAM channels, or on IPTV systems.
Description
In hybrid fibre-coaxial systems, a fiber optic network extending from the operator's head end carries video channels out to a fiber optic node that services up to 2000 end points. Video is then sent via coaxial cable. Note that only a percentage of these homes are actively watching channels at a given time. Rarely are all channels being accessed by the homes in the service group.
In a switched video system, the unwatched channels do not need to be sent.
In US cable systems, equipment in the home sends a channel request signal back to the distribution hub. If a channel is requested, the distribution hub allocates a QAM channel and transmits the channel to the coaxial cable. For this to work, the home equipment must have two-way communication ability. Switched video uses the same mechanisms as video on demand and may be viewed as non-ending video on demand that users share.
Technical
Two-way communication is handled differently between cable and IPTV schemes. IPTV uses Internet communication protocols but requires a different distribution infrastructure. US cable companies elected the less costly approach of upgrading existing infrastructure. In the upgrade approach, various proprietary schemes use specific frequencies for messaging the distribution hub.
For switched video to work on cable systems, digital television users in a subscription group must have devices capable of communicating to the distribution hub in a compatible manner. Unlike other features dependent on two-way communication such as video on demand, the requirement to upgrade all digital se |
https://en.wikipedia.org/wiki/Hamming%287%2C4%29 | In coding theory, Hamming(7,4) is a linear error-correcting code that encodes four bits of data into seven bits by adding three parity bits. It is a member of a larger family of Hamming codes, but the term Hamming code often refers to this specific code that Richard W. Hamming introduced in 1950. At the time, Hamming worked at Bell Telephone Laboratories and was frustrated with the error-prone punched card reader, which is why he started working on error-correcting codes.
The Hamming code adds three additional check bits to every four data bits of the message. Hamming's (7,4) algorithm can correct any single-bit error, or detect all single-bit and two-bit errors. In other words, the minimal Hamming distance between any two correct codewords is 3, and received words can be correctly decoded if they are at a distance of at most one from the codeword that was transmitted by the sender. This means that for transmission medium situations where burst errors do not occur, Hamming's (7,4) code is effective (as the medium would have to be extremely noisy for two out of seven bits to be flipped).
In quantum information, the Hamming (7,4) is used as the base for the Steane code, a type of CSS code used for quantum error correction.
Goal
The goal of the Hamming codes is to create a set of parity bits that overlap so that a single-bit error in a data bit or a parity bit can be detected and corrected. While multiple overlaps can be created, the general method is presented in Hamming codes.
{| class="wikitable"
|-
!Bit # !! 1 !! 2 !! 3 !! 4 !! 5 !! 6 !! 7
|-
!Transmitted bit !! !! !! !! !! !! !!
|-
|
|
|
|
|
|
|
|
|-
|
|
|
|
|
|
|
|
|-
|
|
|
|
|
|
|
|
|}
This table describes which parity bits cover which transmitted bits in the encoded word. For example, p2 provides an even parity for bits 2, 3, 6, and 7. It also details which transmitted bit is covered by which parity bit by reading the column. For example, d1 is covered by p1 and p2 but no |
https://en.wikipedia.org/wiki/E-reader | An e-reader, also called an e-book reader or e-book device, is a mobile electronic device that is designed primarily for the purpose of reading digital e-books and periodicals.
Any device that can display text on a screen may act as an e-reader; however, specialized e-reader devices may optimize portability, readability, and battery life for this purpose. Their main advantage over printed books is portability. This is because an e-reader is capable of holding thousands of books while weighing less than one book, and the convenience provided due to add-on features.
Overview
An e-reader is a device designed as a convenient way to read e-books. It is similar in form factor to a tablet computer, but often features electronic paper ("e-ink") rather than an LCD screen. This yields much longer battery life — the battery can last for several weeks — and better readability, similar to that of paper even in sunlight. Drawbacks of this kind of display include a slow refresh rate and (usually) a grayscale-only display, which makes it unsuitable for sophisticated interactive applications as those found on tablets. The absence of such apps may be perceived as an advantage, as the user may more easily focus on reading. The Sony Librie, released in 2004 and the precursor to the Sony Reader, was the first e-reader to use electronic paper.
Many e-readers can use the internet through Wi-Fi and the built-in software can provide a link to a digital Open Publication Distribution System (OPDS) library or an e-book retailer, allowing the user to buy, borrow, and receive digital e-books. An e-reader may also download e-books from a computer or read them from a memory card. However, the use of memory cards is decreasing as most of the 2010s era e-readers lack a card slot.
History
An idea similar to that of an e-reader is described in a 1930 manifesto written by Bob Brown titled The Readies, which describes "a simple reading machine which I can carry or move around, attach to any old ele |
https://en.wikipedia.org/wiki/Quantum%20satis | Quantum satis (abbreviation q.s. or Q.S.) is a Latin term meaning the amount which is enough. It has its origins as a quantity specification in medicine and pharmacology, where a similar term quantum sufficit ("as much as is sufficient") has been used (abbreviated Q.S.). Quantum satis is also used in the same function in food regulations and food safety laws in the European Community (EC/EU).
The specification of quantum satis for an ingredient essentially means "Add as much of this ingredient as is needed to achieve the desired result, but not more."
In food safety regulations in the EU it is a catch-all restriction for artificial food ingredients (especially food additives) which are harmless enough to have no specific quantity restriction.
It serves to protect consumers from the addition of excessive and unnecessary amounts of such artificial food additives in their foodstuffs and compels producers to:
Introduce minimal additives to food for human consumption
Observe Good Manufacturing Practice
Refrain from wilful consumer deception
For example, European Union directive 94/36/EC (which regulates the use of food colors) explains in Article 2 (7): "In the Annexes to this Directive 'quantum satis' means that no maximum level is specified. However, coloring matters shall be used according to good manufacturing practice at a level not higher than is necessary to achieve the intended purpose and provided that they do not mislead the customer". The words quantum satis are used with reference to a number of substances in the Annexes III and IV to the EU directive 94/36/EC. |
https://en.wikipedia.org/wiki/Point-of-care%20testing | Point-of-care testing (POCT), also called near-patient testing or bedside testing, is defined as medical diagnostic testing at or near the point of care—that is, at the time and place of patient care. This contrasts with the historical pattern in which testing was wholly or mostly confined to the medical laboratory, which entailed sending off specimens away from the point of care and then waiting hours or days to learn the results, during which time care must continue without the desired information.
Technology overview
Point-of-care tests are simple medical tests that can be performed at the bedside. In many cases, the simplicity was not achievable until technology developed not only to make a test possible at all but then also to mask its complexity. For example, various kinds of urine test strips have been available for decades, but portable ultrasonography did not reach the stage of being advanced, affordable, and widespread until the 2000s and 2010s. Today, portable ultrasonography is often viewed as a "simple" test, but there was nothing simple about it until the more complex technology was available. Similarly, pulse oximetry can test arterial oxygen saturation in a quick, simple, noninvasive, affordable way today, but in earlier eras this required an intra-arterial needle puncture and a laboratory test; and rapid diagnostic tests such as malaria antigen detection tests or COVID-19 rapid tests that rely on a state of the art in immunology that did not exist until recent decades. Thus, over decades, testing continues to move toward the point of care more than it formerly had been. A recent survey in five countries (Australia, Belgium, the Netherlands, the UK and the US) indicates that general practitioners / family doctors would like to use more POCTs.
The driving notion behind POCT is to bring the test conveniently and immediately to the patient. This increases the likelihood that the patient, physician, and care team will receive the results quicker, whi |
https://en.wikipedia.org/wiki/Chou%E2%80%93Fasman%20method | The Chou–Fasman method is an empirical technique for the prediction of secondary structures in proteins, originally developed in the 1970s by Peter Y. Chou and Gerald D. Fasman. The method is based on analyses of the relative frequencies of each amino acid in alpha helices, beta sheets, and turns based on known protein structures solved with X-ray crystallography. From these frequencies a set of probability parameters were derived for the appearance of each amino acid in each secondary structure type, and these parameters are used to predict the probability that a given sequence of amino acids would form a helix, a beta strand, or a turn in a protein. The method is at most about 50–60% accurate in identifying correct secondary structures, which is significantly less accurate than the modern machine learning–based techniques.
Amino acid propensities
The original Chou–Fasman parameters found some strong tendencies among individual amino acids to prefer one type of secondary structure over others. Alanine, glutamate, leucine, and methionine were identified as helix formers, while proline and glycine, due to the unique conformational properties of their peptide bonds, commonly end a helix. The original Chou–Fasman parameters were derived from a very small and non-representative sample of protein structures due to the small number of such structures that were known at the time of their original work. These original parameters have since been shown to be unreliable and have been updated from a current dataset, along with modifications to the initial algorithm.
The Chou–Fasman method takes into account only the probability that each individual amino acid will appear in a helix, strand, or turn. Unlike the more complex GOR method, it does not reflect the conditional probabilities of an amino acid to form a particular secondary structure given that its neighbors already possess that structure. This lack of cooperativity increases its computational efficiency but decreases |
https://en.wikipedia.org/wiki/Discharging%20method%20%28discrete%20mathematics%29 | The discharging method is a technique used to prove lemmas in structural graph theory. Discharging is most well known for its central role in the proof of the four color theorem. The discharging method is used to prove that every graph in a certain class contains some subgraph from a specified list. The presence of the desired subgraph is then often used to prove a coloring result.
Most commonly, discharging is applied to planar graphs.
Initially, a charge is assigned to each face and each vertex of the graph.
The charges are assigned so that they sum to a small positive number. During the Discharging Phase the charge at each face or vertex may be redistributed to nearby faces and vertices, as required by a set of discharging rules. However, each discharging rule maintains the sum of the charges. The rules are designed so that after the discharging phase each face or vertex with positive charge lies in one of the desired subgraphs. Since the sum of the charges is positive, some face or vertex must have a positive charge. Many discharging arguments use one of a few standard initial charge functions (these are listed below). Successful application of the discharging method requires creative design of discharging rules.
An example
In 1904, Wernicke introduced the discharging method to prove the following theorem, which was part of an attempt to prove the four color theorem.
Theorem: If a planar graph has minimum degree 5, then it either has an edge
with endpoints both of degree 5 or one with endpoints of degrees 5 and 6.
Proof:
We use , , and to denote the sets of vertices, faces, and edges, respectively.
We call an edge light if its endpoints are both of degree 5 or are of degrees 5 and 6.
Embed the graph in the plane. To prove the theorem, it is sufficient to only consider planar triangulations (because, if it holds on a triangulation, when removing nodes to return to the original graph, neither node on either side of the desired edge can be removed |
https://en.wikipedia.org/wiki/Articular%20capsule%20of%20the%20knee%20joint | The articular capsule of the knee joint is the wide and lax joint capsule of the knee. It is thin in front and at the side, and contains the patella, ligaments, menisci, and bursae of the knee. The capsule consists of an inner synovial membrane, and an outer fibrous membrane separated by fatty deposits anteriorly and posteriorly.
Synovial membrane
Anteriorly, the reflection of the synovial membrane lies on the femur; located at some distance from the cartilage because of the presence of the suprapatellar bursa. Above, the reflection appears lifted from the bone by underlying periosteal connective tissue. In a standing posture, the suprapatellar bursa is seemingly redundant. It is however also referred to as the suprapatellar synovial recess as it gradually unfolds as the knee is flexed; to open up completely when the knee is flexed 130 degrees. The suprapatellar bursa is prevented from being pinched during extension by the articularis genus muscle. On the tibia, the anterior reflection and attachment of the synovial membrane is located near the cartilage.
Anteriorly, the infrapatellar fat pad is inserted below the patella and between the two membranes. It extends from the lower margin of the patella above, to the infrapatellar synovial fold below. With its free upper margin, this fold extends dorsally through the joint space to surround the two cruciate ligaments from the front, thus dividing the surrounding joint space into two chambers. Laterally of this are a pair of alar folds.
Posteriorly, the femoral attachment of the synovial membrane is located at the cartilaginous margin of the lateral and medial femoral condyles, where the joint space has two dorsal extensions. Between these, the synovial membrane passes in front of the anterior and posterior cruciate ligaments, why these ligaments are both intracapsular and extra-articular with their tibial attachment located exactly on the cartilage margin. Both the lateral and medial meniscus are, howev |
https://en.wikipedia.org/wiki/Surface-enhanced%20laser%20desorption/ionization | Surface-enhanced laser desorption/ionization (SELDI) is a soft ionization method in mass spectrometry (MS) used for the analysis of protein mixtures. It is a variation of matrix-assisted laser desorption/ionization (MALDI). In MALDI, the sample is mixed with a matrix material and applied to a metal plate before irradiation by a laser, whereas in SELDI, proteins of interest in a sample become bound to a surface before MS analysis. The sample surface is a key component in the purification, desorption, and ionization of the sample. SELDI is typically used with time-of-flight (TOF) mass spectrometers and is used to detect proteins in tissue samples, blood, urine, or other clinical samples, however, SELDI technology can potentially be used in any application by simply modifying the sample surface.
Sample preparation and instrumentation
SELDI can be seen as a combination of solid-phase chromatography and TOF-MS. The sample is applied to a modified chip surface, which allows for the specific binding of proteins from the sample to the surface. Contaminants and unbound proteins are then washed away. After washing the sample, an energy absorbing matrix, such as sinapinic acid (SPA) or α-Cyano-4-hydroxycinnamic acid (CHCA), is applied to the surface and allowed to crystallize with the sample. Alternatively, the matrix can be attached to the sample surface by covalent modification or adsorption before the sample is applied. The sample is then irradiated by a pulsed laser, causing ablation and desorption of the sample and matrix.
SELDI-TOF-MS
Samples spotted on a SELDI surface are typically analyzed using time-of-flight mass spectrometry. An irradiating laser ionizes peptides from crystals of the sample/matrix mixture. The matrix absorbs the energy of the laser pulse, preventing destruction of the molecule, and transfers charge to the sample molecules, forming ions. The ions are then briefly accelerated through an electric potential and travel down a field-free flight tube |
https://en.wikipedia.org/wiki/Phosphoproteomics | Phosphoproteomics is a branch of proteomics that identifies, catalogs, and characterizes proteins containing a phosphate group as a posttranslational modification. Phosphorylation is a key reversible modification that regulates protein function, subcellular localization, complex formation, degradation of proteins and therefore cell signaling networks. With all of these modification results, it is estimated that between 30%–65% of all proteins may be phosphorylated, some multiple times. Based on statistical estimates from many datasets, 230,000, 156,000 and 40,000 phosphorylation sites should exist in human, mouse, and yeast, respectively.
Compared to expression analysis, phosphoproteomics provides two additional layers of information. First, it provides clues on what protein or pathway might be activated because a change in phosphorylation status almost always reflects a change in protein activity. Second, it indicates what proteins might be potential drug targets as exemplified by the kinase inhibitor Gleevec. While phosphoproteomics will greatly expand knowledge about the numbers and types of phosphoproteins, its greatest promise is the rapid analysis of entire phosphorylation based signalling networks.
Overview
A sample large-scale phosphoproteomic analysis includes cultured cells undergo SILAC encoding; cells are stimulated with factor of interest (e.g. growth factor, hormone); stimulation can occur for various lengths of time for temporal analysis, cells are lysed and enzymatically digested, peptides are separated using ion exchange chromatography; phosphopeptides are enriched using phosphospecific antibodies, immobilized metal affinity chromatography or titanium dioxide (TiO2) chromatography; phosphopeptides are analyzed using mass spectrometry, and peptides are sequenced and analyzed.
Tools and methods
The analysis of the entire complement of phosphorylated proteins in a cell is certainly a feasible option. This is due to the optimization of enric |
https://en.wikipedia.org/wiki/Rule%20of%20three%20%28C%2B%2B%20programming%29 | The rule of three and rule of five are rules of thumb in C++ for the building of exception-safe code and for formalizing rules on resource management. The rules prescribe how the default members of a class should be used to achieve these goals systematically.
Rule of three
The rule of three (also known as the law of the big three or the big three) is a rule of thumb in C++ (prior to C++11) that claims that if a class defines any of the following then it should probably explicitly define all three:
destructor
copy constructor
copy assignment operator
These three functions are special member functions. If one of these functions is used without first being declared by the programmer it will be implicitly implemented by the compiler with the following default semantics:
Destructor – call the destructors of all the object's class-type members
Copy constructor – construct all the object's members from the corresponding members of the copy constructor's argument, calling the copy constructors of the object's class-type members, and doing a plain assignment of all non-class type (e.g., int or pointer) data members
Copy assignment operator – assign all the object's members from the corresponding members of the assignment operator's argument, calling the copy assignment operators of the object's class-type members, and doing a plain assignment of all non-class type (e.g. int or pointer) data members.
The rule of three claims that if one of these had to be defined by the programmer, it means that the compiler-generated version does not fit the needs of the class in one case and it will probably not fit in the other cases either. The term "Rule of three" was coined by Marshall Cline in 1991.
An amendment to this rule is that if the class is designed in such a way that resource acquisition is initialization (RAII) is used for all its (nontrivial) members, the destructor may be left undefined (also known as The Law of The Big Two). A ready-to-go example of this appro |
https://en.wikipedia.org/wiki/NOS%20%28operating%20system%29 | NOS (Network Operating System) is a discontinued operating system with time-sharing capabilities, written by Control Data Corporation in 1975.
NOS ran on the 60-bit CDC 6000 series of mainframe computers and their successors. NOS replaced the earlier CDC Kronos operating system of the 1970s. NOS was intended to be the sole operating system for all CDC machines, a fact CDC promoted heavily. NOS was replaced with NOS/VE on the 64-bit Cyber-180 systems in the mid-1980s.
Version 1 of NOS continued to be updated until about 1981; NOS version 2 was released early 1982.
Time-sharing commands
ACCESS – selects the access subsystem
APL – selects APL programing language
ASCII – select fill 128-character ASCII
ATTACH – links to a permanent file
AUTO – automatically generate five-digit line numbers
BASIC – selects BASIC system
BATCH – selects the batch system
BEGIN – starts processing of CCL procedure (control language file)
BINARY – selects binary input mode
BRIEF – suppresses headers
BYE – log off the system
CALL – starts processing KCL procedure file (control language before CCL)
CATLIST – lists user's permanent files
CHANGE – changes parameters of a permanent file
CHARGE – set charge number and project number
CLEAR – releases all local files
CONVERT – converts character sets
(CR) – Carriage Return – requests terminal status if it is the first thing on a line
CSET – selects the terminal character-set mode
DAYFILE – lists a record of the user's activity
DEBUG – activates or terminates CYBER interactive Debug
DEFINE – create a direct-access permanent file
DIAL – sends a one-line message to another terminal
EDIT – Selects the text editor
ENQUIRE – Requests the current job status
EXECUTE – selects the Execute subsystem
FORTRAN – selects the FORTRAN subsystem (FORTRAN 5)
FTNTS – Selects the FORTRAN Extended Version 4 compiler (CDC's enhanced version of FORTRAN 4)
FULL – Selects full-duplex mode
GET – gets a copy of a permanent file
GOODBYE – same |
https://en.wikipedia.org/wiki/Nikto%20%28vulnerability%20scanner%29 | Nikto is a free software command-line vulnerability scanner that scans web servers for dangerous files/CGIs, outdated server software and other problems. It performs generic and server type specific checks. It also captures and prints any cookies received. The Nikto code itself is free software, but the data files it uses to drive the program are not. Version 1.00 was released December 27, 2001.
Features
Nikto can detect over 6700 potentially dangerous files/CGIs, checks for outdated versions of over 1250 servers, and version specific problems on over 270 servers. It also checks for server configuration items such as the presence of multiple index files and HTTP server options, and will attempt to identify installed web servers and software. Scan items and plugins are frequently updated and can be automatically updated.
Variations
There are some variations of Nikto, one of which is MacNikto. MacNikto is an AppleScript GUI shell script wrapper built in Apple's Xcode and Interface Builder, released under the terms of the GPL. It provides easy access to a subset of the features available in the command-line version, installed along with the MacNikto application. |
https://en.wikipedia.org/wiki/Journal%20of%20Food%20Science | The Journal of Food Science is a peer-reviewed scientific journal that was established in 1936 and is published by John Wiley & Sons on behalf of the Institute of Food Technologists in Chicago, Illinois. From 1996 to 2005, it was ranked eighth among impact in scientific journals publishing food science and technology.
History
The journal was founded in 1936 as Food Research with Fred W. Tanner (University of Illinois at Urbana–Champaign) as editor in chief. Published bimonthly by Garrard Press, it was a publication that dealt with food science and technology research. The first issue had nine articles in it. By the end of 1936, 55 papers were published.
In 1950, Food Research was purchased by the Institute of Food Technologists and Zoltan I. Kertesz was named Editor-In-Chief in 1951. Kertesz and most of his successors also served as editors of Food Technology, the institute's magazine founded in 1947. He was replaced by Martin S. Peterson in July 1952, who served until December 1960.
George F. Stewart (University of California, Davis) took over in January 1961, renaming Food Research to it current name. He was succeeded by Walter M. Urbain in July 1966. Ernest J. Briskey edited from June 1970 until January 1971.
In January 1971, all of the applied research articles were shifted from Food Technology to the Journal of Food Science and Bernard J. Liska became editor-in-chief until 1981. He was succeeded by Aaron E. Wasserman who stepped down in 1990.
Robert E. Berry became editor-in-chief in 1990 and stayed until 1998. From 1996 the journal was sectioned by discipline (food chemistry, food engineering, food microbiology, nutrition, and sensory analysis). Subsequently, Owen R. Fennema was editor until September 2003, during which time the journal's publication frequency increased from six issues a year to its current nine issues. The current Editor-in-Chief of JFS is Edward Allen Foegeding, a distinguished professor of food chemistry at North Carolina State Univer |
https://en.wikipedia.org/wiki/Conservation%20form | Conservation form or Eulerian form refers to an arrangement of an equation or system of equations, usually representing a hyperbolic system, that emphasizes that a property represented is conserved, i.e. a type of continuity equation. The term is usually used in the context of continuum mechanics.
General form
Equations in conservation form take the form
for any conserved quantity , with a suitable function . An equation of this form can be transformed into an integral equation
using the divergence theorem. The integral equation states that the change rate of the integral of the quantity over an arbitrary control volume is given by the flux through the boundary of the control volume, with being the outer surface normal through the boundary. is neither produced nor consumed inside of and is hence conserved. A typical choice for is , with velocity , meaning that the quantity flows with a given velocity field.
The integral form of such equations is usually the physically more natural formulation, and the differential equation arises from differentiation. Since the integral equation can also have non-differentiable solutions, the equality of both formulations can break down in some cases, leading to weak solutions and severe numerical difficulties in simulations of such equations.
Example
An example of a set of equations written in conservation form are the Euler equations of fluid flow:
Each of these represents the conservation of mass, momentum and energy, respectively.
See also
Conservation law
Lagrangian and Eulerian specification of the flow field
Further reading
Randall J. LeVeque: Finite Volume Methods for Hyperbolic Problems. Cambridge University Press, Cambridge 2002, (Cambridge Texts in Applied Mathematics).
Algebra
Conservation equations |
https://en.wikipedia.org/wiki/History%20of%20the%20battery | Batteries provided the primary source of electricity before the development of electric generators and electrical grids around the end of the 19th century. Successive improvements in battery technology facilitated major electrical advances, from early scientific studies to the rise of telegraphs and telephones, eventually leading to portable computers, mobile phones, electric cars, and many other electrical devices.
Students and engineers developed several commercially important types of battery. "Wet cells" were open containers that held liquid electrolyte and metallic electrodes. When the electrodes were completely consumed, the wet cell was renewed by replacing the electrodes and electrolyte. Open containers are unsuitable for mobile or portable use. Wet cells were used commercially in the telegraph and telephone systems. Early electric cars used semi-sealed wet cells.
One important classification for batteries is by their life cycle. "Primary" batteries can produce current as soon as assembled, but once the active elements are consumed, they cannot be electrically recharged. The development of the lead-acid battery and subsequent "secondary" or "chargeable" types allowed energy to be restored to the cell, extending the life of permanently assembled cells. The introduction of nickel and lithium based batteries in the latter half of the 20th century made the development of innumerable portable electronic devices feasible, from powerful flashlights to mobile phones. Very large stationary batteries find some applications in grid energy storage, helping to stabilize electric power distribution networks.
Invention
From the mid 18th century on, before there were batteries, experimenters used Leyden jars to store electrical charge. As an early form of capacitor, Leyden jars, unlike electrochemical cells, stored their charge physically and would release it all at once. Many experimenters took to hooking several Leyden jars together to create a stronger charge and |
https://en.wikipedia.org/wiki/Multiplication%20theorem | In mathematics, the multiplication theorem is a certain type of identity obeyed by many special functions related to the gamma function. For the explicit case of the gamma function, the identity is a product of values; thus the name. The various relations all stem from the same underlying principle; that is, the relation for one special function can be derived from that for the others, and is simply a manifestation of the same identity in different guises.
Finite characteristic
The multiplication theorem takes two common forms. In the first case, a finite number of terms are added or multiplied to give the relation. In the second case, an infinite number of terms are added or multiplied. The finite form typically occurs only for the gamma and related functions, for which the identity follows from a p-adic relation over a finite field. For example, the multiplication theorem for the gamma function follows from the Chowla–Selberg formula, which follows from the theory of complex multiplication. The infinite sums are much more common, and follow from characteristic zero relations on the hypergeometric series.
The following tabulates the various appearances of the multiplication theorem for finite characteristic; the characteristic zero relations are given further down. In all cases, n and k are non-negative integers. For the special case of n = 2, the theorem is commonly referred to as the duplication formula.
Gamma function–Legendre formula
The duplication formula and the multiplication theorem for the gamma function are the prototypical examples. The duplication formula for the gamma function is
It is also called the Legendre duplication formula or Legendre relation, in honor of Adrien-Marie Legendre. The multiplication theorem is
for integer k ≥ 1, and is sometimes called Gauss's multiplication formula, in honour of Carl Friedrich Gauss. The multiplication theorem for the gamma functions can be understood to be a special case, for the trivial Dirichlet charac |
https://en.wikipedia.org/wiki/PHI-base | https://canto.phi-base.org/
The Pathogen-Host Interactions database (PHI-base) is a biological database that contains curated information on genes experimentally proven to affect the outcome of pathogen-host interactions. The database is maintained by researchers at Rothamsted Research, together with external collaborators since 2005.
Since April 2017 PHI-base is part of ELIXIR, the European life-science infrastructure for biological information via its ELIXIR-UK node.
Background
The Pathogen-Host Interactions database was developed to utilise effectively the growing number of verified genes that mediate an organism's ability to cause disease and / or to trigger host responses.
The web-accessible database catalogues experimentally verified pathogenicity, virulence and effector genes from bacterial, fungal and oomycete pathogens which infect animal, plant and fungal hosts. PHI-base is the first on-line resource devoted to the identification and presentation of information on fungal and oomycete pathogenicity genes and their host interactions. As such, PHI-base aims to be a resource for the discovery of candidate targets in medically and agronomically important fungal and oomycete pathogens for intervention with synthetic chemistries and natural products (fungicides).
Each entry in PHI-base is curated by domain experts and supported by strong experimental evidence (gene disruption experiments) as well as literature references in which the experiments are described. Each gene in PHI-base is presented with its nucleotide and deduced amino acid sequence as well as a detailed structured description of the predicted protein's function during the host infection process. To facilitate data interoperability, genes are annotated using controlled vocabularies (Gene Ontology terms, EC Numbers, etc.), and links to other external data sources such as UniProt, EMBL and the NCBI taxonomy services.
Current developments
Version 4.15 (May 2, 2023) of PHI-base provides informa |
https://en.wikipedia.org/wiki/Resolvent%20set | In linear algebra and operator theory, the resolvent set of a linear operator is a set of complex numbers for which the operator is in some sense "well-behaved". The resolvent set plays an important role in the resolvent formalism.
Definitions
Let X be a Banach space and let be a linear operator with domain . Let id denote the identity operator on X. For any , let
A complex number is said to be a regular value if the following three statements are true:
is injective, that is, the corestriction of to its image has an inverse ;
is a bounded linear operator;
is defined on a dense subspace of X, that is, has dense range.
The resolvent set of L is the set of all regular values of L:
The spectrum is the complement of the resolvent set:
The spectrum can be further decomposed into the point/discrete spectrum (where condition 1 fails), the continuous spectrum (where conditions 1 and 3 hold but condition 2 fails) and the residual/compression spectrum (where condition 1 holds but condition 3 fails).
If is a closed operator, then so is each , and condition 3 may be replaced by requiring that is surjective.
Properties
The resolvent set of a bounded linear operator L is an open set.
More generally, the resolvent set of a densely defined closed unbounded operator is an open set. |
https://en.wikipedia.org/wiki/Laplacian%20smoothing | Laplacian smoothing is an algorithm to smooth a polygonal mesh. For each vertex in a mesh, a new position is chosen based on local information (such as the position of neighbours) and the vertex is moved there. In the case that a mesh is topologically a rectangular grid (that is, each internal vertex is connected to four neighbours) then this operation produces the Laplacian of the mesh.
More formally, the smoothing operation may be described per-vertex as:
Where is the number of adjacent vertices to node , is the position of the -th adjacent vertex and is the new position for node .
See also
Tutte embedding, an embedding of a planar mesh in which each vertex is already at the average of its neighbours' positions |
https://en.wikipedia.org/wiki/Systematic%20code | In coding theory, a systematic code is any error-correcting code in which the input data are embedded in the encoded output. Conversely, in a non-systematic code the output does not contain the input symbols.
Systematic codes have the advantage that the parity data can simply be appended to the source block, and receivers do not need to recover the original source symbols if received correctly – this is useful for example if error-correction coding is combined with a hash function for quickly determining the correctness of the received source symbols, or in cases where errors occur in erasures and a received symbol is thus always correct. Furthermore, for engineering purposes such as synchronization and monitoring, it is desirable to get reasonable good estimates of the received source symbols without going through the lengthy decoding process which may be carried out at a remote site at a later time.
Properties
Every non-systematic linear code can be transformed into a systematic code with essentially the same properties (i.e., minimum distance).
Because of the advantages cited above, linear error-correcting codes are therefore generally implemented as systematic codes. However, for certain decoding algorithms such as sequential decoding or maximum-likelihood decoding, a non-systematic structure can increase performance in terms of undetected decoding error probability when the minimum free distance of the code is larger.
For a systematic linear code, the generator matrix, , can always be written as , where is the identity matrix of size .
Examples
Checksums and hash functions, combined with the input data, can be viewed as systematic error-detecting codes.
Linear codes are usually implemented as systematic error-correcting codes (e.g., Reed-Solomon codes in CDs).
Convolutional codes are implemented as either systematic or non-systematic codes. Non-systematic convolutional codes can provide better performance under maximum-likelihood (Viterbi) decoding.
In |
https://en.wikipedia.org/wiki/Mr.%20Do%27s%20Wild%20Ride | Mr. Do's Wild Ride is a platform game released in 1984 as the third in Universal's Mr. Do! arcade video game series. An MSX version was published in 1985.
Gameplay
Mr. Do!'s scenario is a roller coaster, and the object is to reach the top. As the cars (and eventually other objects) speed around the track, the player must escape by using a super speed button, or by climbing up small ladders scattered about the track to dodge the hazards. Two icons at the end of the level range from cakes to EXTRA letters or diamonds change upon collecting cherries at the top of each letter. The game is timed, and the timer ticks faster when the super speed button is held down. Collision with a roller coaster car or another object is fatal, knocking Mr. Do! off the coaster and costing a life.
After the sixth level is completed, the game cycles back to the first with various obstacles and/or more roller coaster cars to avoid.
Reception
In Japan, Game Machine listed Mr. Do's Wild Ride on their April 15, 1984 issue as being the sixth most-successful table arcade unit of the month.
Legacy
Ocean Software published a clone for the Commodore 64, Amstrad CPC, and ZX Spectrum as Kong Strikes Back! It incorporates cosmetic aspects of Donkey Kong. |
https://en.wikipedia.org/wiki/Balo%20concentric%20sclerosis | Baló's concentric sclerosis is a disease in which the white matter of the brain appears damaged in concentric layers, leaving the axis cylinder intact. It was described by József Mátyás Baló who initially named it "leuko-encephalitis periaxialis concentrica" from the previous definition, and it is currently considered one of the borderline forms of multiple sclerosis.
Baló's concentric sclerosis is a demyelinating disease similar to standard multiple sclerosis, but with the particularity that the demyelinated tissues form concentric layers. Scientists used to believe that the prognosis was similar to Marburg multiple sclerosis, but now they know that patients can survive, or even have spontaneous remission and asymptomatic cases.
The concentric ring appearance is not specific to Baló's MS. Concentric lesions have also been reported in patients with neuromyelitis optica, standard MS, progressive multifocal leukoencephalopathy, cerebral autosomal dominant arteriopathy with subcortical infarcts, leukoencephalopathy, concomitant active hepatitis C and human herpes virus 6.
Pathophysiology
The lesions of the Baló's sclerosis belong to the MS lesion pattern III (distal oligodendrogliopathy). Balo concentric sclerosis is now believed to be a variant of pattern III multiple sclerosis and probably due to metabolic problems.
The Baló lesions show veins at their center, like those of MS, some suggestive of microhemorrhages or small ectatic venules. Unlike MS, no cortical gray matter lesions appear.
Theoretical models
According with Dr. Lucchinetti investigations, in Baló's concentric sclerosis, the rings may be caused by a physiological hypoxia (similar to that caused by some toxins or viruses) in the lesion, which is in turn countered by expression of stress proteins at the border. This expression and counter-expression forms rings of preserved tissue within the lesion and rings of demyelinated tissue just beyond where the previous attack had induced the protective str |
https://en.wikipedia.org/wiki/Stokesian%20dynamics | Stokesian dynamics
is a solution technique for the Langevin equation, which is the relevant form of Newton's 2nd law for a Brownian particle. The method treats the suspended particles in a discrete sense while the continuum approximation remains valid for the surrounding fluid, i.e., the suspended particles are generally assumed to be significantly larger than the molecules of the solvent. The particles then interact through hydrodynamic forces transmitted via the continuum fluid, and when the particle Reynolds number is small, these forces are determined through the linear Stokes equations (hence the name of the method). In addition, the method can also resolve non-hydrodynamic forces, such as Brownian forces, arising from the fluctuating motion of the fluid, and interparticle or external forces. Stokesian Dynamics can thus be applied to a variety of problems, including sedimentation, diffusion and rheology, and it aims to provide the same level of understanding for multiphase particulate systems as molecular dynamics does for statistical properties of matter. For rigid particles of radius suspended in an incompressible Newtonian fluid of viscosity and density , the motion of the fluid is governed by the Navier–Stokes equations, while the motion of the particles is described by the coupled equation of motion:
In the above equation is the particle translational/rotational velocity
vector of dimension 6N. is the hydrodynamic force, i.e., force exerted by the fluid on the particle due to relative motion between them. is the stochastic Brownian force due to thermal motion of fluid particles. is the deterministic nonhydrodynamic force, which may be almost any form of interparticle or external force, e.g. electrostatic repulsion between like charged particles. Brownian dynamics is one of the popular techniques of solving the Langevin equation, but the hydrodynamic interaction in Brownian dynamics is highly simplified and normally includes only the isolated body |
https://en.wikipedia.org/wiki/Computer%20architecture%20simulator | A computer architecture simulator is a program that simulates the execution of computer architecture.
Computer architecture simulators are used for the following purposes:
Lowering cost by evaluating hardware designs without building physical hardware systems.
Enabling access to unobtainable hardware.
Increasing the precision and volume of computer performance data.
Introducing abilities that are not normally possible on real hardware such as running code backwards when an error is detected or running in faster-than-real time.
Categories
Computer architecture simulators can be classified into many different categories depending on the context.
Scope: Microarchitecture simulators model the microprocessor and its components. Full-system simulators also model the processor, memory systems, and I/O devices.
Detail: Functional simulators, such as instruction set simulators, achieve the same function as modeled components. They can be simulated faster if timing is not considered. Timing simulators are functional simulators that also reproduce timing. Timing simulators can be further categorized into digital cycle-accurate and analog sub-cycle simulators.
Workload: Trace-driven simulators (also called event-driven simulators) react to pre-recorded streams of instructions with some fixed input. Execution-driven simulators allow dynamic change of instructions to be executed depending on different input data.
Full-system simulators
A full-system simulator is execution-driven architecture simulation at such a level of detail that complete software stacks from real systems can run on the simulator without any modification. A full system simulator provides virtual hardware that is independent of the nature of the host computer. The full-system model typically includes processor cores, peripheral devices, memories, interconnection buses, and network connections. Emulators are full system simulators that imitate obsolete hardware instead of under development hardware.
|
https://en.wikipedia.org/wiki/Ethmoidal%20notch | The ethmoidal notch separates the two orbital plates; it is quadrilateral, and filled, in the articulated skull, by the cribriform plate of the ethmoid.
The margins of the notch present several half-cells which, when united with corresponding half-cells on the upper surface of the ethmoid, complete the ethmoidal sinuses. |
https://en.wikipedia.org/wiki/Namco%20System%20246 | The Namco System 246 is a development of the Sony PlayStation 2 technology as a basis for an arcade system board. It was released in December 2000 on its first game Bloody Roar 3. Like the Sega NAOMI, it is widely licensed for use by other manufacturers. Games such as Battle Gear 3 and Capcom Fighting Evolution are examples of System 246-based arcade games that are not Namco products.
Specifications
Main CPU: MIPS III R5900-based "Emotion Engine", 64-bit RISC operating at 294.912 MHz (Overclocking to 299 MHz on System 256), with 128-bit SIMD capabilities
Sub CPU: MIPS II R3000A IOP with cache at 33.8688 MHz (Unlike the PSXCPU)
System memory: 32 MB RIMM 3200 32-bit dual-channel (2x 16-bit) RDRAM (Direct Rambus DRAM) @ 400 MHz, 3.2 GB/s peak bandwidth
Graphics: "Graphics Synthesizer" operating at 147.456 MHz
Graphics memory: 4MB eDRAM (8MB on System 256)
Sound: "SPU1+SPU2"
Media: CD-ROM, DVD-ROM, Hard Disk Drive, 64 MBit 3.3V NAND FlashROM Security Dongle
Namco System 256 is an upgraded version of System 246, but the upgrades are unknown (more VRAM and faster CPU speeds likely).
Namco Super System 256 is the same as regular Namco System 256 but it has the gun board integrated, though this variant was only used in Time Crisis 4.
Namco System 147 is similar to 246 but does not use a DVD-ROM drive, instead it has ROM chips on the system board.
List of System 246 / System 256 / System Super 256 / System 147 games
Released
Ace Driver 3: Final Turn (Namco, 2008)
Battle Gear 3 (Taito, 2002)
Battle Gear 3 Tuned (Taito, 2003)
Bloody Roar 3 (Namco / 8ing / Raizing, 2000)
Capcom Fighting Evolution (Capcom, 2004)
Chou Dragon Ball Z (Banpresto, 2005)
Cobra: The Arcade (Namco, 2005)
Dragon Chronicles - Legend of The Master Ark (Namco, 2003)
Dragon Chronicles Online (Namco, 2004)
Druaga Online - The Story of Aon (Namco, 2006)
Fate/Unlimited Codes (Capcom / Type-Moon / Cavia / 8ing, 2008)
Getchu Play! Tottoko Hamutaro (Banpresto, 2007)
Gundam Seed: Rengou |
https://en.wikipedia.org/wiki/Moonsound | Moonsound is the name of a sound card released for the MSX home-computer system at the Tilburg Computer Fair in 1995. It was designed by electronic engineer Henrik Gilvad and produced by Sunrise Swiss on a semi-hobby basis.
It arrived after the US branch of Microsoft abandoned the MSX system, instead focusing on the IBM PC. The name originates from the Moonblaster software that was written for people to use the hardware plug-in synthesizer.
Overview
Based on the Yamaha YMF278 (OPL4) sound chip, it is capable of 18 channels of FM synthesis as well as 24 channels of 12 and 16 bit sample-based synthesis. A 2 MB instrument ROM containing multisampled instruments was unusual for its time. From the factory it came equipped with one 128 KB SRAM chip for user samples.
History
Two generations were made. The first is a small size PCB without a box. Later, a larger size PCB which fit into an MSX cartridge was available. The later version had room for two sample SRAM chips resulting in 1 MB of compressed user samples.
Sound effects
Sound effects like chorus, delay and reverb are omitted due to cost, size and usability reasons. The Yamaha effect chip requires its own specialised memory and effect routing is basic. All 18 FM channels and 24 channels of sample-based sound shares the same effect setting. Creative step-time sequencer programmers made pseudo effects like chorus, reverb and delay by overdubbing or using dedicated channels to repeat notes with delay and stereo panning. This is effective but quickly reduces the musical complexity possible.
Specifications
Moonsound version 1.0 had one socket for user sample RAM.
Moonsound version 1.1 and 1.2 had two sockets for up to 1 MB SRAM.
Some hackers and modders found out how to stack two additional SRAM chips resulting in 2 MB of SRAM.
Being based on the OPL4 chip, The FM registers of Moonsound are compatible with the OPL, OPL2 and OPL3 chips. The MSX-AUDIO contains a chip which is similar to and also compatible with th |
https://en.wikipedia.org/wiki/Etching%20%28microfabrication%29 | Etching is used in microfabrication to chemically remove layers from the surface of a wafer during manufacturing. Etching is a critically important process module in fabrication, and every wafer undergoes many etching steps before it is complete.
For many etch steps, part of the wafer is protected from the etchant by a "masking" material which resists etching. In some cases, the masking material is a photoresist which has been patterned using photolithography. Other situations require a more durable mask, such as silicon nitride.
Etching media and technology
The two fundamental types of etchants are liquid-phase ("wet") and plasma-phase ("dry"). Each of these exists in several varieties.
Wet etching
The first etching processes used liquid-phase ("wet") etchants. This process is now largely outdated but was used up until the late 1980s when it was superseded by dry plasma etching. The wafer can be immersed in a bath of etchant, which must be agitated to achieve good process control. For instance, buffered hydrofluoric acid (BHF) is used commonly to etch silicon dioxide over a silicon substrate.
Different specialized etchants can be used to characterize the surface etched.
Wet etchants are usually isotropic, which leads to a large bias when etching thick films. They also require the disposal of large amounts of toxic waste. For these reasons, they are seldom used in state-of-the-art processes. However, the photographic developer used for photoresist resembles wet etching.
As an alternative to immersion, single wafer machines use the Bernoulli principle to employ a gas (usually, pure nitrogen) to cushion and protect one side of the wafer while etchant is applied to the other side. It can be done to either the front side or back side. The etch chemistry is dispensed on the top side when in the machine and the bottom side is not affected. This etching method is particularly effective just before "backend" processing (BEOL), where wafers are normally very much |
https://en.wikipedia.org/wiki/Lacrimal%20tubercle | The lateral margin of the groove of the frontal process of the maxilla is named the anterior lacrimal crest, and is continuous below with the orbital margin; at its junction with the orbital surface is a small tubercle, the lacrimal tubercle, which serves as a guide to the position of the lacrimal sac. |
https://en.wikipedia.org/wiki/Lateral%20palpebral%20raphe | The lateral palpebral raphe is a ligamentous band near the eye. Its existence is contentious, and many sources describe it as the continuation of nearby muscles. It is formed from the lateral ends of the orbicularis oculi muscle. It connects the orbicularis oculi muscle, the frontosphenoidal process of the zygomatic bone, and the tarsi of the eyelids.
Structure
The lateral palpebral raphe is formed from the lateral ends of the orbicularis oculi muscle. It may also be formed from the pretarsal muscles of the eyelids. It is attached to the margin of the frontosphenoidal process of the zygomatic bone. It passes towards the midline to the lateral commissure of the eyelids. Here, it divides into two slips, which are attached to the margins of the respective tarsi of the eyelids.
The lateral palpebral ligament has a tensile strength of around 12 newtons.
Relations
The lateral palpebral raphe is a much weaker structure than the medial palpebral ligament on the other side of the eyelids.
Variation
The lateral palpebral raphe may be absent in some people. If it is not present, it is replaced with muscular fibres of orbicularis oculi muscle. It is often very hard to identify as a distinct anatomical feature. Some sources claim that it does not exist.
See also
Orbicularis oculi muscle
Lateral palpebral commissure |
https://en.wikipedia.org/wiki/Orbital%20lamina%20of%20ethmoid%20bone | The orbital lamina of ethmoid bone (or lamina papyracea or orbital lamina) is a smooth, oblong, paper-thin bone plate which forms the lateral wall of the labyrinth of the ethmoid bone. It covers the middle and posterior ethmoidal cells, and forms a large part of the medial wall of the orbit.
It articulates above with the orbital plate of the frontal bone, below with the maxilla and the orbital process of palatine bone, in front with the lacrimal, and behind with the sphenoid.
Its name lamina papyracea is an appropriate description, as this part of the ethmoid bone is paper-thin and fractures easily. A fracture here could cause entrapment of the medial rectus muscle.
Additional Images |
https://en.wikipedia.org/wiki/Sybil%20attack | A Sybil attack is a type of attack on a computer network service in which an attacker subverts the service's reputation system by creating a large number of pseudonymous identities and uses them to gain a disproportionately large influence. It is named after the subject of the book Sybil, a case study of a woman diagnosed with dissociative identity disorder. The name was suggested in or before 2002 by Brian Zill at Microsoft Research. The term pseudospoofing had previously been coined by L. Detweiler on the Cypherpunks mailing list and used in the literature on peer-to-peer systems for the same class of attacks prior to 2002, but this term did not gain as much influence as "Sybil attack".
Description
The Sybil attack in computer security is an attack wherein a reputation system is subverted by creating multiple identities. A reputation system's vulnerability to a Sybil attack depends on how cheaply identities can be generated, the degree to which the reputation system accepts inputs from entities that do not have a chain of trust linking them to a trusted entity, and whether the reputation system treats all entities identically. , evidence showed that large-scale Sybil attacks could be carried out in a very cheap and efficient way in extant realistic systems such as BitTorrent Mainline DHT.
An entity on a peer-to-peer network is a piece of software that has access to local resources. An entity advertises itself on the peer-to-peer network by presenting an identity. More than one identity can correspond to a single entity. In other words, the mapping of identities to entities is many to one. Entities in peer-to-peer networks use multiple identities for purposes of redundancy, resource sharing, reliability and integrity. In peer-to-peer networks, the identity is used as an abstraction so that a remote entity can be aware of identities without necessarily knowing the correspondence of identities to local entities. By default, each distinct identity is usually assumed |
https://en.wikipedia.org/wiki/Hess%20diagram | A Hess diagram plots the relative density of occurrence of stars at differing color–magnitude positions of the Hertzsprung–Russell diagram for a given galaxy or resolved stellar population. The diagram is named after R. Hess who originated it in 1924. Its use dates back to at least 1948.
Hess diagrams are widely used in the study of discrete resolved stellar systems in and around the Milky Way - specifically, in the analysis of globular clusters, satellite galaxies, and stellar streams.
See also
Color-color diagram |
https://en.wikipedia.org/wiki/Stirling%20numbers%20and%20exponential%20generating%20functions%20in%20symbolic%20combinatorics | The use of exponential generating functions (EGFs) to study the properties of Stirling numbers is a classical exercise in combinatorial mathematics and possibly the canonical example of how symbolic combinatorics is used. It also illustrates the parallels in the construction of these two types of numbers, lending support to the binomial-style notation that is used for them.
This article uses the coefficient extraction operator for formal power series, as well as the (labelled) operators (for cycles) and (for sets) on combinatorial classes, which are explained on the page for symbolic combinatorics. Given a combinatorial class, the cycle operator creates the class obtained by placing objects from the source class along a cycle of some length, where cyclical symmetries are taken into account, and the set operator creates the class obtained by placing objects from the source class in a set (symmetries from the symmetric group, i.e. an "unstructured bag".) The two combinatorial classes (shown without additional markers) are
permutations (for unsigned Stirling numbers of the first kind):
and
set partitions into non-empty subsets (for Stirling numbers of the second kind):
where is the singleton class.
Warning: The notation used here for the Stirling numbers is not that of the Wikipedia articles on Stirling numbers; square brackets denote the signed Stirling numbers here.
Stirling numbers of the first kind
The unsigned Stirling numbers of the first kind count the number of permutations of [n] with k cycles. A permutation is a set of cycles, and hence the set of permutations is given by
where the singleton marks cycles. This decomposition is examined in some detail on the page on the statistics of random permutations.
Translating to generating functions we obtain the mixed generating function of the unsigned Stirling numbers of the first kind:
Now the signed Stirling numbers of the first kind are obtained from the unsigned ones through the relation
Hence |
https://en.wikipedia.org/wiki/Scalded%20milk | Scalded milk is dairy milk that has been heated to . At this temperature, bacteria are killed, enzymes in the milk are destroyed, and many of the proteins are denatured. Since most milk sold today is pasteurized, which accomplishes the first two goals, milk is typically scalded to increase its temperature, or to change the consistency or other cooking interactions due to the denaturing of proteins.
During scalding, a milk watcher (a cooking utensil) may be used to prevent both boiling over and scorching (burning) of the milk.
Uses
Béchamel sauce
Scalded milk is called for in the original recipes for Béchamel sauce, as adding hot liquid, including milk, to a roux was thought less likely to make a lumpy sauce or one tasting of raw flour.
Bread
Scalded and cooled milk is used in bread and other yeast doughs, as pasteurization does not kill all bacteria, and with the wild yeasts that may also be present, these can alter the texture and flavor. Recipes old enough to have been based on hand-milked, slowly cooled, unpasteurized milk specify scalded milk with much more justification, and modern cookbooks tend to maintain the tradition. In addition, scalding milk improves the rise due to inhibition of bread rise by certain undenatured milk proteins.
Yogurt
Scalded milk is used in yogurt to make the proteins unfold, and to make sure that all organisms that could outcompete the yogurt culture's bacteria are killed. In traditional yogurt making, as done in the Eastern Mediterranean and Near East, the milk is often heated in flat pans until reduced to about half. Whatever the effect of scalding on milk protein may be, it is mainly this concentrating that reduces whey separation. Modern commercial processors use dried or concentrated milk or vegetable gums and gelatins such as pectin, carrageenan, or agar ("vegetable gelatin") to prevent whey separation in yogurt.
Other uses
Café au lait and baked milk use scalded milk, ryazhenka uses baked milk.
Scalded and cooled mil |
https://en.wikipedia.org/wiki/Dual%20quaternion | In mathematics, the dual quaternions are an 8-dimensional real algebra isomorphic to the tensor product of the quaternions and the dual numbers. Thus, they may be constructed in the same way as the quaternions, except using dual numbers instead of real numbers as coefficients. A dual quaternion can be represented in the form , where A and B are ordinary quaternions and ε is the dual unit, which satisfies and commutes with every element of the algebra.
Unlike quaternions, the dual quaternions do not form a division algebra.
In mechanics, the dual quaternions are applied as a number system to represent rigid transformations in three dimensions. Since the space of dual quaternions is 8-dimensional and a rigid transformation has six real degrees of freedom, three for translations and three for rotations, dual quaternions obeying two algebraic constraints are used in this application. Since unit quaternions are subject to two algebraic constraints, unit quaternions are standard to represent rigid transformations.
Similar to the way that rotations in 3D space can be represented by quaternions of unit length, rigid motions in 3D space can be represented by dual quaternions of unit length. This fact is used in theoretical kinematics (see McCarthy), and in applications to 3D computer graphics, robotics and computer vision. Polynomials with coefficients given by (non-zero real norm) dual quaternions have also been used in the context of mechanical linkages design.
History
W. R. Hamilton introduced quaternions in 1843, and by 1873 W. K. Clifford obtained a broad generalization of these numbers that he called biquaternions, which is an example of what is now called a Clifford algebra.
In 1898 Alexander McAulay used Ω with Ω2 = 0 to generate the dual quaternion algebra. However, his terminology of "octonions" did not stick as today's octonions are another algebra.
In Russia, Aleksandr Kotelnikov developed dual vectors and dual quaternions for use in the study of mec |
https://en.wikipedia.org/wiki/P1%20phage | P1 is a temperate bacteriophage that infects Escherichia coli and some other bacteria. When undergoing a lysogenic cycle the phage genome exists as a plasmid in the bacterium unlike other phages (e.g. the lambda phage) that integrate into the host DNA. P1 has an icosahedral head containing the DNA attached to a contractile tail with six tail fibers.
The P1 phage has gained research interest because it can be used to transfer DNA from one bacterial cell to another in a process known as transduction. As it replicates during its lytic cycle it captures fragments of the host chromosome. If the resulting viral particles are used to infect a different host the captured DNA fragments can be integrated into the new host's genome. This method of in vivo genetic engineering was widely used for many years and is still used today, though to a lesser extent. P1 can also be used to create the P1-derived artificial chromosome cloning vector which can carry relatively large fragments of DNA. P1 encodes a site-specific recombinase, Cre, that is widely used to carry out cell-specific or time-specific DNA recombination by flanking the target DNA with loxP sites (see Cre-Lox recombination).
Morphology
The virion is similar in structure to the T4 phage but simpler. It has an icosahedral head containing the genome attached at one vertex to the tail. The tail has a tube surrounded by a contractile sheath. It ends in a base plate with six tail fibres. The tail fibres are involved in attaching to the host and providing specificity.
Genome
The genome of the P1 phage is moderately large, around 93Kbp in length (compared to the genomes of e.g. T4 - 169Kbp, lambda - 48Kbp and Ff - 6.4Kbp). In the viral particle it is in the form of a linear double stranded DNA molecule. Once inserted into the host it circularizes and replicates as a plasmid.
In the viral particle the DNA molecule is longer (110Kbp) than the actual length of the genome. It is created by cutting an appropriately sized fr |
https://en.wikipedia.org/wiki/Life%20Sciences%20Greenhouse%20of%20Central%20Pennsylvania | Life Sciences Greenhouse of Central Pennsylvania (LSGPA) is a biotechnology initiative and non-profit organization based in Harrisburg, Pennsylvania. It was founded in 2001. It focuses on in the advancement of life sciences through technology to improve the healthcare and economic opportunities of Pennsylvanians.
Background
The initiative began in 2001, funded from the state's settlement with the tobacco industry. Other life sciences greenhouses in Philadelphia and Pittsburgh also received seed money from the settlement. LSGPA partners with a range of institutions, including local research universities, colleges, medical centers, economic development agencies and companies of various sizes to identify needs and opportunities. It then works to help transfer technologies, develop new companies, provide support for existing companies (particularly those seeking to expand or relocate), and ensure that the infrastructure to support a thriving life sciences industry keeps pace with development.
Research areas
Central Pennsylvania has three large research universities which contribute to the initiative. Collectively, these three institutions attract more than $600 million in sponsored research funding annually. They are:
Lehigh University, located in Bethlehem, Pennsylvania
Penn State University, located in State College, Pennsylvania
Penn State Hershey Medical Center, located in Hershey, Pennsylvania |
https://en.wikipedia.org/wiki/Pre-math%20skills | Pre-math skills (referred to in British English as pre-maths skills) are math skills learned by preschoolers and kindergarten students, including learning to count numbers (usually from 1 to 10 but occasionally including 0), learning the proper sequencing of numbers, learning to determine which shapes are bigger or smaller, and learning to count objects on a screen or book. Pre-math skills are also tied into literacy skills to learn the correct pronunciations of numbers. |
https://en.wikipedia.org/wiki/Peppercoin | Peppercoin is a cryptographic system for processing micropayments. Peppercoin Inc. was a company that offers services based on the peppercoin method.
The peppercoin system was developed by Silvio Micali and Ron Rivest and first presented at the RSA Conference in 2002 (although it had not yet been named.) The core idea is to bill one randomly selected transaction a lump sum of money rather than bill each transaction a small amount. It uses "universal aggregation", which means that it aggregates transactions over users, merchants as well as payment service providers. The random selection is cryptographically secure—it cannot be influenced by any of the parties. It is claimed to reduce the transaction cost per dollar from 27 cents to "well below 10 cents."
Peppercoin, Inc. was a privately held company founded in late 2001 by Micali and Rivest based in Waltham, MA. It has secured about $15M in venture capital in two rounds of funding. Its services have seen modest adoption. Peppercoin collects 5-9% of transaction cost from the merchant. Peppercoin, Inc. was bought out in 2007 by Chockstone for an undisclosed amount. |
https://en.wikipedia.org/wiki/GOR%20method | The GOR method (short for Garnier–Osguthorpe–Robson) is an information theory-based method for the prediction of secondary structures in proteins. It was developed in the late 1970s shortly after the simpler Chou–Fasman method. Like Chou–Fasman, the GOR method is based on probability parameters derived from empirical studies of known protein tertiary structures solved by X-ray crystallography. However, unlike Chou–Fasman, the GOR method takes into account not only the propensities of individual amino acids to form particular secondary structures, but also the conditional probability of the amino acid to form a secondary structure given that its immediate neighbors have already formed that structure. The method is therefore essentially Bayesian in its analysis.
Method
The GOR method analyzes sequences to predict alpha helix, beta sheet, turn, or random coil secondary structure at each position based on 17-amino-acid sequence windows. The original description of the method included four scoring matrices of size 17×20, where the columns correspond to the log-odds score, which reflects the probability of finding a given amino acid at each position in the 17-residue sequence. The four matrices reflect the probabilities of the central, ninth amino acid being in a helical, sheet, turn, or coil conformation. In subsequent revisions to the method, the turn matrix was eliminated due to the high variability of sequences in turn regions (particularly over such a large window). The method was considered as best requiring at least four contiguous residues to score as alpha helices to classify the region as helical, and at least two contiguous residues for a beta sheet.
Algorithm
The mathematics and algorithm of the GOR method were based on an earlier series of studies by Robson and colleagues reported mainly in the Journal of Molecular Biology and The Biochemical Journal. The latter describes the information theoretic expansions in terms of conditional information measures. The |
https://en.wikipedia.org/wiki/Ben%20Franklin%20effect | The Ben Franklin effect is a psychological phenomenon in which people like someone more after doing a favour for them. An explanation for this is cognitive dissonance. People reason that they help others because they like them, even if they do not, because their minds struggle to maintain logical consistency between their actions and perceptions.
The Benjamin Franklin effect, in other words, is the result of one's concept of self coming under attack. Every person develops a persona, and that persona persists because inconsistencies in one's personal narrative get rewritten, redacted, and misinterpreted.
Franklin's observation of effect
Benjamin Franklin, after whom the effect is named, quoted what he described as an "old maxim" in his autobiography: "He that has once done you a kindness will be more ready to do you another, than he whom you yourself have obliged."
Franklin explains how he dealt with the animosity of a rival legislator when he served in the Pennsylvania Assembly in the 18th century:
Research
A study of the effect was done by Jecker and Landy in 1969, in which students were invited to take part in a Q&A competition run by the researcher in which they could win sums of money. After this competition was over, one-third of the students who had "won" were approached by the researcher, who asked them to return the money on the grounds that he had used his own funds to pay the winners and was running short of money now; another third were asked by a secretary to return the money because it was from the psychology department and funds were low; another third were not at all approached. All three groups were then asked how much they liked the researcher. The second group liked him the least, the first group the mostsuggesting that a refund request by an intermediary had decreased their liking, while a direct request had increased their liking.
In 1971, University of North Carolina psychologists John Schopler and John Compere carried out the following ex |
https://en.wikipedia.org/wiki/Pegylated%20interferon | Pegylated interferon (PEG-IFN) is a class of medication that includes three different drugs as of 2012:
Pegylated interferon-alpha-2a
Pegylated interferon-alpha-2b
Pegylated interferon beta-1a
In these formulations, Polyethylene glycol (PEG) is added to make interferon last longer in the body. They are used to treat both hepatitis B, hepatitis C and multiple sclerosis.
Pegylated interferon is contraindicated in patients with hyperbilirubinaemia. |
https://en.wikipedia.org/wiki/A%E1%B9%A3%E1%B9%AD%C4%81dhy%C4%81y%C4%AB | The () is a grammar that describes a form of an early Indo-Aryan language: Sanskrit.
Authored by Sanskrit philologist and scholar Pāṇini and dated to around 500 BCE, it describes the language as current in his time, specifically the dialect and register of an élite of model speakers, referred to by Pāṇini himself as śiṣṭa. The work also accounts both for some features specific to the older Vedic form of the language, as well as certain dialectal features current in the author's time.
The Aṣṭādhyāyī employs a derivational system to describe the language, where real speech is derived from posited abstract utterances formed by means of affixes added to bases under certain conditions.
The Aṣṭādhyāyī is supplemented by three ancillary texts: akṣarasamāmnāya, dhātupāṭha and gaṇapāṭha.
Etymology
Aṣṭādhyāyī is made of two words aṣṭa-, 'eight' and adhyāya-, 'chapter', thus meaning eight-chaptered, or 'the book of eight chapters'.
Background
Grammatical tradition
By 1000 BCE, a large body of hymns composed in the oldest attested form of the Proto-Indo-Aryan language had been consolidated into the Rigveda, which formed the canonical basis of the Vedic religion, being transmitted from generation to generation entirely orally.
In the course of the following centuries, as the popular speech evolved, growing concern among the guardians of the Vedic religion that the hymns be passed on without 'corruption' led to the rise of a vigorous, sophisticated grammatical tradition involving the study of linguistic analysis, in particular phonetics alongside grammar. The high point of this centuries-long endeavour was Pāṇini's Aṣṭādhyāyī, which eclipsed all others before him.
While not the first, the Aṣṭādhyāyī is the oldest linguistic and grammar text, and one of the oldest Sanskrit texts, surviving in its entirety. Pāṇini refers to older texts such as the Unādisūtra, Dhātupāṭha, and Gaṇapātha but some of these have only survived in part.
Arrangement
The Aṣṭādhyāyī consists |
https://en.wikipedia.org/wiki/Fluorescein%20diacetate%20hydrolysis | Fluorescein diacetate (FDA) hydrolysis assays can be used to measure the enzyme activity of microbes in a sample. A bright yellow-green glow is produced and is strongest when enzymatic activity is greatest. This can be quantified using a spectrofluorometer or a spectrophotometer.
Applications
FDA hydrolysis is often used to measure activity in soil and compost samples; however, it may not give an accurate reading if microbes with lower activity phases, such as esterases, cleave the fluorescein first.
It is also used in combination with propidium iodide (PI) to determine viability in eukaryotic cells. Living cells will actively convert the non-fluorescent FDA into the green fluorescent compound fluorescein, a sign of viability; while nucleus of membrane-compromised cells will fluoresce red, a sign of cell death. Currently FDA/PI staining is the standard assessment of human pancreatic islet viability with suitability for transplantation when viability score is above 70%.
Preparation
FDA stock solution is prepared by dissolving 5 mg of fluorescein diacetate in 1 ml acetone, and sucrose may be added for live cell viability testing. FDA stain must be kept in the dark at 4°C or it will spoil. |
https://en.wikipedia.org/wiki/Cinnamon%20basil | Cinnamon basil is a type of basil (Ocimum basilicum). The term "cinnamon basil" can refer to a number of different varieties of basil, including as a synonym for Thai basil (O. basilicum var. thyrsiflora), as a particular cultivar of Thai basil, and as a separate cultivar in its own right (i.e., O. basilicum 'Cinnamon'). This article discusses the latter type.
Description
Cinnamon basil, also known as Mexican spice basil, has a spicy, fragrant aroma and flavor. It contains methyl cinnamate, giving it a flavor reminiscent of cinnamon. Cinnamon basil has somewhat narrow, slightly serrated, dark green, shiny leaves with reddish-purple veins, which can resemble certain types of mint, and produces small, pink flowers from July to September. Its stems are dark purple. Cinnamon basil grows to 18–30 inches tall.
Cultivation
Cinnamon basil is an easy-to-grow herb. It requires six to eight hours of bright sunlight per day. Although it is often grown as an annual, it is a perennial in USDA plant hardiness zones 9–11. Cinnamon basil is sometimes planted near tomatoes and roses to discourage pests such as whiteflies.
Uses
Cinnamon basil is used in teas and baked goods such as cookies and pies. It is also used in pastas, salads, jellies, and vinegars. Outside the kitchen, cinnamon basil is used in dried arrangements and as a potpourri.
Space
Cinnamon basil was taken into space by the Space Shuttle Endeavour during STS-118 and grown in an experiment in low Earth orbit on the International Space Station. |
https://en.wikipedia.org/wiki/Blowing%20agent | A blowing agent is a substance which is capable of producing a cellular structure via a foaming process in a variety of materials that undergo hardening or phase transition, such as polymers, plastics, and metals. They are typically applied when the blown material is in a liquid stage. The cellular structure in a matrix reduces density, increasing thermal and acoustic insulation, while increasing relative stiffness of the original polymer.
Blowing agents (also known as 'pneumatogens') or related mechanisms to create holes in a matrix producing cellular materials, have been classified as follows:
Physical blowing agents include CFCs (however, these are ozone depletants, banned by the Montreal Protocol of 1987), HCFCs (replaced CFCs, but are still ozone depletants, therefore being phased out), hydrocarbons (e.g. pentane, isopentane, cyclopentane), and liquid CO2. The bubble/foam-making process is irreversible and endothermic, i.e. it needs heat (e.g. from a melt process or the chemical exotherm due to cross-linking), to volatilize a liquid blowing agent. However, on cooling the blowing agent will condense, i.e. a reversible process.
Chemical blowing agents include isocyanate and water for polyurethane, azodicarbonamide for vinyl, hydrazine and other nitrogen-based materials for thermoplastic and elastomeric foams, and sodium bicarbonate for thermoplastic foams. Gaseous products and other byproducts are formed by a chemical reaction of the chemical blowing agent, promoted by the heat of the foam production process or a reacting polymer's exothermic heat. Since the blowing reaction occurs forming low molecular weight compounds acting as the blowing gas, additional exothermic heat is also released. Powdered titanium hydride is used as a foaming agent in the production of metal foams, as it decomposes to form hydrogen gas and titanium at elevated temperatures. Zirconium(II) hydride is used for the same purpose. Once formed the low molecular weight compounds will neve |
https://en.wikipedia.org/wiki/MacCharlie | The MacCharlie was a hardware add-on for the Apple Macintosh that was made by Dayna Communications. It allowed users to run DOS software for the IBM PC by clipping a unit onto the chassis of the Macintosh 128K, and included a keyboard extender to provide the function keys and numeric keypad that are absent from Apple's original keyboard. The name refers to an IBM PC advertising campaign of the time featuring Charlie Chaplin's "Little Tramp" character.
The clip-on unit sits to the side of the Mac and, like the contemporary Amiga Sidecar, contains essentially a complete IBM PC compatible with an 8088 processor, 256 KB of RAM (expandable to 640 KB) and a single 5.25" floppy disk drive that stores 360 KB. A second floppy drive could be added.
While running DOS software using MacCharlie, users could still access the Macintosh menu bar and desk accessories. However, the DOS environment, which ran in a window, was text-only and did not permit Macintosh applications to run concurrently while in use. MacCharlie used the Mac as a terminal, performing all DOS processing itself, and sent video data over a relatively slow serial link to the Mac for display. This slowness, coupled with the declining prices of real IBM PC compatibles, contributed to the short market life of the MacCharlie.
See also
Amiga Sidecar |
https://en.wikipedia.org/wiki/Video%20game%20localization | Video game localization (American English), or video game localisation (British English; see spelling differences), is the process of preparing a video game for a market outside of where it was originally published. The game's name, art assets, packaging, manuals, and cultural and legal differences are typically altered.
Before localization, producers consider economic factors such as potential foreign profit. Most official localizations are done by the game's developers or a third-party translation company. Nevertheless, fan localizations are also popular.
Localization is largely inconsistent between platforms, engines and companies due to its recency. Localizers intend to create an experience like the original game, with discretion to the localization audience. Localizations are considered to have failed if it is confusing or difficult to understand; this may break the player's immersion.
History
Since the beginning of video game history, video games have been localized. One of the first widely popular video games, Pac-Man was localized from Japanese. The original transliteration of the Japanese title would be "Puck-Man," but the decision was made to change the name when the game was imported to the United States out of fear that the word 'Puck' would be vandalized into an obscenity. In addition, the names of the ghosts were originally based on colors - roughly translating to "Reddie," "Pinky," "Bluey," and "Slowly." Rather than translate these names exactly, they were renamed to Blinky, Pinky, Inky, and Clyde. This choice maintained the odd-man-out style of the original names without adhering to their exact meaning. This is an early example of a change in cultural context.
Early localization had one main concern. Due to the small memory size of the NES and SNES cartridges many translated text strings were too long. Ted Woolsey, translator of Final Fantasy VI, recounts having to continually cut down the English text due to limited capacity.
Early video gam |
https://en.wikipedia.org/wiki/Food%20Science%20and%20Technology%20Abstracts | FSTA, also known as FSTA – Food Science and Technology Abstracts, is produced by IFIS Publishing.
FSTA is a bibliographic abstracting and indexing (A&I) database of scientific and technological research and information relating to food, beverages, and nutrition. It contains over 1,400,000 indexed records, with full-text links where available.
The database is used by researchers, industry practitioners, and university students.
Coverage
In addition to over 5,475 active and historical journals, FSTA indexes books, trade publications, reviews, conference proceedings, reports, patents, and standards, producing 22,675 sources overall. Updated weekly, its records are indexed against IFIS' thesaurus, which contains over 12,346 food science keywords, curated and structured into food-centric hierarchies.
With records dating back to 1969, FSTA contains information sources in 29 languages, sourced from publishers in over 60 countries.
Coverage includes all major commodities in the food and beverage industry, related applied and pure sciences, pet foods, food psychology, food economics, food safety, and more.
Online access
FSTA can be accessed through EBSCOhost, Ovid, Proquest Dialog, STN and Web of Science.
See also
Google Scholar
List of academic databases and search engines
Lists of academic journals
List of open-access journals
List of scientific journals |
https://en.wikipedia.org/wiki/Electronic%20tongue | The electronic tongue is an instrument that measures and compares tastes. As per the IUPAC technical report, an “electronic tongue” as analytical instrument including an array of non-selective chemical sensors with partial specificity to different solution components and an appropriate pattern recognition instrument, capable to recognize quantitative and qualitative compositions of simple and complex solutions
Chemical compounds responsible for taste are detected by human taste receptors. Similarly, the multi-electrode sensors of electronic instruments detect the same dissolved organic and inorganic compounds. Like human receptors, each sensor has a spectrum of reactions different from the other. The information given by each sensor is complementary, and the combination of all sensors' results generates a unique fingerprint. Most of the detection thresholds of sensors are similar to or better than human receptors.
In the biological mechanism, taste signals are transduced by nerves in the brain into electric signals. E-tongue sensors process is similar: they generate electric signals as voltammetric and potentiometric variations.
Taste quality perception and recognition are based on the building or recognition of activated sensory nerve patterns by the brain and the taste fingerprint of the product. This step is achieved by the e-tongue's statistical software, which interprets the sensor data into taste patterns.
Operation
Liquid samples are directly analyzed without any preparation, whereas solids require a preliminary dissolution before measurement. Reference electrode and sensors are dipped in a beaker containing a test solution. A voltage is applied between each sensor and a reference electrode, and a measurable current response results that is consistent with the Cottrell equation. This current response is a result of oxidizing reactions that take place in the solution due to the voltage difference and can be amplified through catalytic surface treatments. |
https://en.wikipedia.org/wiki/XG%20Station | The ASUS XG Station is a device designed to provide external graphics processing power to laptops.
It connects to a laptop though an ExpressCard slot. It requires a separate monitor as well as its own power source.
It includes USB 2.0 ports, a headphone jack, and a large knob used to control various settings such as overclocking. The screen tells the user information such as the GPU's clock and memory speeds, fan speeds, temperature, master volume, and FPS. From available photos it would appear that it will also provide dual DVI output connectors.
Since XG station uses the Express Card interface, actual bandwidth available to the card will be approximately PCI-E 1.1 x1 bandwidth.
The XG Station was featured at the 2007 Consumer Electronics Show.
The XG Station was scheduled to be released at the beginning of Q2 2007. A full package will include the XG Station graphics docking station, one ASUS EN7900GS graphics card and assorted accessories according to Asus News. The EN7900GS graphics card is an Nvidia GeForce 7900 according to an article from the Inquirer, and can be swapped out for another one. A January 2008 publication renewed speculation that the device was approaching production, and the XG Station reached limited release in May 2008.
In early 2008, the XG Station was only made available in Australia. It consisted of an NVIDIA GeForce 8600GT with 256MB DDR3 for approximately A$375. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.