text
stringlengths
0
473k
[SOURCE: https://he.wikipedia.org/wiki/ืคื•ืจื˜ืœ:ื™ืฉืจืืœ] | [TOKENS: 4720]
ืคื•ืจื˜ืœ:ื™ืฉืจืืœ ืจืขื ื•ืŸ ื”ืคื•ืจื˜ืœ ื›ื™ืฆื“ ืื•ื›ืœ ืœืขื–ื•ืจ? ืžื“ื™ื ืช ื™ืฉืจืืœ ื”ืžื•ื“ืจื ื™ืช ื ื•ืกื“ื” ื‘-1948 (ื”' ื‘ืื™ื™ืจ ืชืฉ"ื—), ื›ืฉืœื•ืฉ ืฉื ื™ื ืœืื—ืจ ื”ืฉื•ืื”. ื”ืงืžืชื” ืœื•ื•ืชื” ื‘ืžืœื—ืžื” ืขื ื”ืขืจื‘ื™ื ื”ืžืงื•ืžื™ื™ื ื•ืืจืฆื•ืช ืขืจื‘ ื”ืฉื›ื ื•ืช ืœื”. ื”ืกื›ืกื•ืš ื”ืขืจื‘ื™-ื™ืฉืจืืœื™ ืงื™ื™ื ื•ืฉืจื™ืจ ืขื“ ื”ื™ื•ื, ืœืžืจื•ืช ืžืืžืฆื™ ืคื™ื•ืก ื•ืฉืœื•ื ืฉื ืขืฉื• ื‘ืžื”ืœืš ื”ืฉื ื™ื. ื‘ืฉืœ ื”ืกื›ืกื•ืš, ื‘ื™ืฉืจืืœ ื™ืฉ ื’ื™ื•ืก ื—ื•ื‘ื” ืœืฆื‘ื ื”ื’ื ื” ืœื™ืฉืจืืœ, ื”ื ื—ืฉื‘ ืœืฆื‘ื ื”...
========================================
[SOURCE: https://he.wikipedia.org/wiki/%D7%95%D7%99%D7%A7%D7%99%D7%A4%D7%93%D7%99%D7%94:%D7%9E%D7%96%D7%A0%D7%95%D7%9F] | [TOKENS: 64263]
ืชื•ื›ืŸ ืขื ื™ื™ื ื™ื ื•ื™ืงื™ืคื“ื™ื”:ืžื–ื ื•ืŸ ื”ืžื–ื ื•ืŸ ืžืฉืžืฉ ืœื“ื™ื•ื ื™ ืžื“ื™ื ื™ื•ืช ื•ื“ื™ื•ื ื™ื ื›ืœืœื™ื™ื ืฉืื™ืŸ ืœื”ื ืžืงื•ื ืžืชืื™ื ืื—ืจ. ื‘ืขื ื™ื™ื ื™ื ืื—ืจื™ื ื™ืฉ ืœืคื ื•ืช ืœื“ืคื™ื ื”ื‘ืื™ื: [ ืขืจื™ื›ื” ] โ†“ ืžืขื‘ืจ ืœืชื—ืชื™ืช ื”ื“ืฃโ†“ ืžืขื‘ืจ ืœืชื—ืชื™ืช ื”ื“ืฃ ื”ื›ืชื™ื‘ "ื ืฆื™" ื‘ืฉืžื•ืช ืขืจื›ื™ื ืžืฆื‘ ื˜ื™ืคื•ืœ: ื‘ื“ื™ื•ืŸ ื‘ื›ืžื” ื•ื›ืžื” ืขืจื›ื™ื ืžืฆื•ื™ ื”ื›ืชื™ื‘ "ื ืฆื™" (ืื• "ื ืฆื™ื"), ื›ืชื™ื‘ ืžื™ื•ืฉืŸ ืฉื›ื‘ืจ ืื™ื ื• ืชืงื ื™. ื”ื›ืชื™ื‘ ื”ืชืงื ื™ ื›ื™ื•ื ื”ื•ื "ื ื™ืฆื™", ...
========================================
[SOURCE: https://he.wikipedia.org/wiki/ืžื•ื ื—] | [TOKENS: 1428]
ืชื•ื›ืŸ ืขื ื™ื™ื ื™ื ื˜ืจืžื™ื ื•ืœื•ื’ื™ื” ื˜ืจืžื™ื ื•ืœื•ื’ื™ื” ืื• ืžื™ื ื•ื— ื”ื™ื ืชื—ื•ื ืžื—ืงืจ ืชืื•ืจื˜ื™ ืฉืขื ื™ื™ื ื• ืžื•ื ื—ื™ื ื•ืฉื™ืžื•ืฉื™ื”ื. "ืžื•ื ื—ื™ื", ื‘ืžืกื’ืจืช ื”ื˜ืจืžื™ื ื•ืœื•ื’ื™ื”, ื”ื ืžื™ืœื™ื, ืื• ืžื™ืœื™ื-ืžื•ืจื›ื‘ื•ืช, ื”ืžืฉืžืฉื•ืช ื‘ื”ืงืฉืจ ืžืกื•ื™ื, ืœื”ื‘ื“ื™ืœ ืžืŸ ื”ืฉื™ืžื•ืฉ ื”ืฆืจ ื”ืžืฆื•ื™ ื‘ืžื™ืœื”, ื”ืžืฆื™ื™ืŸ ืื•ืกืฃ ื”ื’ื“ืจื•ืช ื™ื™ื—ื•ื“ื™ื•ืช ื”ืžืฉืžืฉื•ืช ื‘ืžืกื’ืจืช ืคืจืงื˜ื™ืงื” ืžืกื•ื™ืžืช. ืžื—ืงืจื™ื ื‘ื˜ืจืžื™ื ื•ืœื•ื’ื™ื” ืขื•ืกืงื™ื ื‘ื™ืŸ ื”ื™ืชืจ ื‘ืฉืืœืช ื”ื™ื•ื•ืฆืจื•ืชื ืฉืœ...
========================================
[SOURCE: https://he.wikipedia.org/wiki/ืกืค_ืคื™ื•ื ื˜ืง] | [TOKENS: 3493]
ืชื•ื›ืŸ ืขื ื™ื™ื ื™ื ืกืค ืคื™ื•ื ื˜ืง ื™ื•ื–ืฃ ืขืžื ื•ืืœ ื”ื•ื‘ืจื˜ื•ืก ืคื™ื•ื ื˜ืง (ื‘ื’ืจืžื ื™ืช: Josef Emanuel Hubertus Piontek;โ€ 5 ื‘ืžืจืฅ 1940 โ€“ 18 ื‘ืคื‘ืจื•ืืจ 2026), ื•ื‘ืงื™ืฆื•ืจ ืกืค ืคื™ื•ื ื˜ืง (Sepp Piontek), ื”ื™ื” ื›ื“ื•ืจื’ืœืŸ ื•ืžืืžืŸ ื’ืจืžื ื™. ืคื™ื•ื ื˜ืง ืฉื™ื—ืง ื›ืžืขื˜ ืœืื•ืจืš ื›ืœ ื”ืงืจื™ื™ืจื” ืฉืœื• ื‘ืขืžื“ืช ื”ื‘ืœื ื‘ื•ื•ืจื“ืจ ื‘ืจืžืŸ, ื–ื›ื” ืขืžื” ื‘ืชื•ืืจ ืืœื™ืคื•ืช ื”ื‘ื•ื ื“ืกืœื™ื’ื” ื‘ืขื•ื ืช 1964/65, ื•ืขืจืš ืฉืฉ ื”ื•ืคืขื•ืช ื‘ื ื‘ื—ืจืช ื’ืจืžื ื™...
========================================
[SOURCE: https://he.wikipedia.org/wiki/ื’'ืกื™_ื’'ืงืกื•ืŸ] | [TOKENS: 3675]
ืชื•ื›ืŸ ืขื ื™ื™ื ื™ื ื’'ืกื™ ื’'ืงืกื•ืŸ ื’'ืกื™ ืœื•ืื™ืก ื’'ืงืกื•ืŸ (ื‘ืื ื’ืœื™ืช: Jesse Louis Jackson;โ€ 8 ื‘ืื•ืงื˜ื•ื‘ืจ 1941 โ€“ 17 ื‘ืคื‘ืจื•ืืจ 2026) ื”ื™ื” ื›ื•ืžืจ, ืคื•ืœื™ื˜ื™ืงืื™ ื•ืคืขื™ืœ ื‘ืชื ื•ืขื” ืœื–ื›ื•ื™ื•ืช ื”ืื–ืจื— ื‘ืืจืฆื•ืช ื”ื‘ืจื™ืช. ื”ื•ื ื”ื™ื” ืžื ื”ื™ื’ ื‘ื•ืœื˜ ืฉืœ ื”ืฉืžืืœ ื”ื ื•ืฆืจื™ ื•ื”ืชืžื•ื“ื“ ืขืœ ืžื•ืขืžื“ื•ืช ื”ืžืคืœื’ื” ื”ื“ืžื•ืงืจื˜ื™ืช ืœื ืฉื™ืื•ืช ืืจืฆื•ืช ื”ื‘ืจื™ืช ื‘-1984 ื•ื‘-1988. ื‘ื™ื•ื’ืจืคื™ื” ื’'ืงืกื•ืŸ ื ื•ืœื“ ื‘ืฉื ื’'ืกื™ ืœื•ืื™ืก ื‘ืจื ืก ื‘...
========================================
[SOURCE: https://he.wikipedia.org/wiki/ื™ื”ื“ื•ืช] | [TOKENS: 17867]
ืชื•ื›ืŸ ืขื ื™ื™ื ื™ื ื™ื”ื“ื•ืช ื”ื™ื”ื“ื•ืช ื”ื™ื ืจืฆืฃ ื”ืžืกื•ืจืช ื•ื”ื”ืžืฉื›ื™ื•ืช ื”ื“ืชื™ืช, ื”ืชืจื‘ื•ืชื™ืช ื•ื”ืžืฉืคื˜ื™ืช ื”ืงื•ืœืงื˜ื™ื‘ื™ืช ืฉืœ ื”ื™ื”ื•ื“ื™ื. ืžื•ืจืฉืช ื ืจื—ื‘ืช ื–ื• ื›ื•ืœืœืช ื ืจื˜ื™ื‘ื™ื ื”ื™ืกื˜ื•ืจื™ื•ื’ืจืคื™ื™ื ืžืฉื•ืชืคื™ื ื•ืืœืžื ื˜ื™ื ืฉืœ ื“ืช, ืœืื•ื ื•ืชืจื‘ื•ืช, ืฉื”ืชืคืชื—ื• ื‘ืงืจื‘ ืขื ื™ืฉืจืืœ ื”ื—ืœ ืžืจืืฉื™ืช ื”ืชื’ื‘ืฉื•ืชื• ื‘ืื–ื•ืจ ื”ืœื‘ื ื˜ ื‘ืฉืœื”ื™ ื”ืืœืฃ ื”ืฉื ื™ ืœืคื ื”"ืก. ื ื™ืชืŸ ืœื ืงื•ื˜ ื‘ืžื•ื ื— ื™ื”ื“ื•ืช ื›ืœืคื™ ื ื•ืฉืื™ ืžื•ืจืฉืช ื–ื• ื”ื—ืœ ืžื™ืžื™ ื‘ื™ืช ืฉื ื™, ืœ...
========================================
[SOURCE: https://he.wikipedia.org/wiki/ืชื ื•ืขืช_ื”ื—ืกื™ื“ื•ืช] | [TOKENS: 47822]
ืชื•ื›ืŸ ืขื ื™ื™ื ื™ื ืชื ื•ืขืช ื”ื—ืกื™ื“ื•ืช ืœืฉืขืจ ืœื ื•ืฉืื™ื, ืื™ืฉื™ื ื•ืžืืžืจื™ื ื‘ืชื ื•ืขืช ื”ื—ืกื™ื“ื•ืช, ืจืื• ืคื•ืจื˜ืœ ืชื ื•ืขืช ื”ื—ืกื™ื“ื•ืช ืชื ื•ืขืช ื”ื—ืกื™ื“ื•ืช ื”ื™ื ืชื ื•ืขื” ืจื•ื—ื ื™ืช ื•ื—ื‘ืจืชื™ืช ื™ื”ื•ื“ื™ืช ืฉืงืžื” ื‘ืืžืฆืข ื”ืžืื” ื”-18 ื‘ืžืขืจื‘ ืื•ืงืจืื™ื ื” ืฉืœ ื”ื™ื•ื, ื•ื”ืชืคืฉื˜ื” ื‘ืžื”ื™ืจื•ืช ื‘ื™ื”ื“ื•ืช ืžื–ืจื— ืื™ืจื•ืคื”. ื”ื‘ืขืœ ืฉื ื˜ื•ื‘ ื ื—ืฉื‘ ืœืžื—ื•ืœืœื” ืฉืœ ื”ื—ืกื™ื“ื•ืช. ืชืœืžื™ื“ื™ื•, ื‘ืจืืฉื•ืช ื”ืžื’ื™ื“ ืžืžื–ืจื™ื˜ืฉ, ืคื™ืชื—ื• ืืช ืชื•ืจืชื” ื•ื”ืคื™ืฆื•ื”. ื”ื’ื•ืชื” ...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#References] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#Pure_statistical_models] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#Large_language_models] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://he.wikipedia.org/wiki/ื—ืžืฆื”] | [TOKENS: 4392]
ืชื•ื›ืŸ ืขื ื™ื™ื ื™ื ื—ืžืฆื” ื—ึดืžึฐืฆึธื” ืชืจื‘ื•ืชื™ืช (ืฉื ืžื“ืขื™: Cicer arietinum), ื”ืžื•ื›ืจืช ื™ื•ืชืจ ื‘ืฉืžื” ื”ืขืจื‘ื™ ื—ื•ืžื•ืก (ืชืขืชื™ืง ืžื“ื•ื™ืง: ื—ึปืžึปึผืฅ), ื”ื™ื ืฆืžื— ืžืžืฉืคื—ืช ื”ืงื˜ื ื™ื•ืช. ื’ื•ื‘ื”ื• ืฉืœ ื”ืฆืžื— ื”ื•ื 20 - 50 ืก"ืž, ื•ื”ื•ื ื‘ืขืœ ืขืœื™ื ื ื•ืฆืชื™ื™ื ืงื˜ื ื™ื ื”ื ืžืฆืื™ื ืžืฉื ื™ ืฆื“ื™ ื”ื’ื‘ืขื•ืœ. ื‘ืชืจืžื™ืœ ื–ืจืข ืื—ื“ ืขื“ ืฉืœื•ืฉื” ื–ืจืขื™ ื—ืžืฆื”. ื”ืคืจื—ื™ื ืœื‘ื ื™ื ืื• ื›ื—ื•ืœื™ื ืื“ืžื“ืžื™ื. ื—ืžืฆื” ื–ืงื•ืงื” ืœืืงืœื™ื ื—ื ื•ืœืžืขืœื” ืžึพ4...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Natural_language] | [TOKENS: 623]
Contents Natural language A natural language or ordinary language is any spoken language or signed language used organically in a human community, first emerging without conscious premeditation and subject to: replication across generations of people in the community, regional expansion or contraction, and gradual inte...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#cite_note-3] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#cite_note-Semantic_parsing_as_machine_translation-4] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#cite_note-5] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#cite_note-2] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Grammar_induction] | [TOKENS: 1686]
Contents Grammar induction Grammar induction (or grammatical inference) is the process in machine learning of learning a formal grammar (usually as a collection of re-write rules or productions or alternatively as a finite-state machine or automaton of some kind) from a set of observations, thus constructing a model wh...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Special:RecentChangesLinked/Language_model] | [TOKENS: 57]
Related changes Enter a page name to see changes on pages linked to or from that page. (To see members of a category, enter Category:Name of category). Changes to pages on your Watchlist are shown in bold with a green bullet. See more at Help:Related changes.
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#cite_note-6] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Recurrent_neural_network] | [TOKENS: 6352]
Contents Recurrent neural network In artificial neural networks, recurrent neural networks (RNNs) are designed for processing sequential data, such as text, speech, and time series, where the order of elements is important. Unlike feedforward neural networks, which process inputs independently, RNNs utilize recurrent c...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Text_corpus] | [TOKENS: 345]
Contents Text corpus In linguistics and natural language processing, a corpus (pl.: corpora) or text corpus is a dataset, consisting of natively digital and older, digitalized, language resources, either annotated or unannotated. Annotated, they have been used in corpus linguistics for statistical hypothesis testing, c...
========================================
[SOURCE: https://en.wikipedia.org/w/index.php?title=Word_n-gram_language_model&action=edit] | [TOKENS: 1451]
Editing Word n-gram language model Copy and paste: โ€“ โ€” ยฐ โ€ฒ โ€ณ โ‰ˆ โ‰  โ‰ค โ‰ฅ ยฑ โˆ’ ร— รท โ† โ†’ ยท ยง Cite your sources: <ref></ref> {{}} {{{}}} | [] [[]] [[Category:]] #REDIRECT [[]] &nbsp; <s></s> <sup></sup> <sub></sub> <code></code> <pre></pre> <blockquote></blockquote> <ref></ref> <ref name="" /> {{Reflist}} <references /> <includ...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#cite_note-14] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Language_model#cite_note-12] | [TOKENS: 1793]
Contents Language model A language model is a computational model that predicts sequences in natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimiz...
========================================
[SOURCE: https://en.wikipedia.org/wiki/Principle_of_maximum_entropy] | [TOKENS: 5066]
Contents Principle of maximum entropy The principle of maximum entropy states that, among all probability distributions consistent with a given set of constraints (such as normalization or specified expectation values), the distribution that maximizes Shannon entropy should be selected. This yields the least committal ...
========================================