Spaces:
Sleeping
Sleeping
| /* | |
| * Copyright (c) 2010 Arc90 Inc | |
| * | |
| * Licensed under the Apache License, Version 2.0 (the "License"); | |
| * you may not use this file except in compliance with the License. | |
| * You may obtain a copy of the License at | |
| * | |
| * http://www.apache.org/licenses/LICENSE-2.0 | |
| * | |
| * Unless required by applicable law or agreed to in writing, software | |
| * distributed under the License is distributed on an "AS IS" BASIS, | |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
| * See the License for the specific language governing permissions and | |
| * limitations under the License. | |
| */ | |
| /* | |
| * This code is heavily based on Arc90's readability.js (1.7.1) script | |
| * available at: http://code.google.com/p/arc90labs-readability | |
| */ | |
| /** | |
| * Public constructor. | |
| * @param {HTMLDocument} doc The document to parse. | |
| * @param {Object} options The options object. | |
| */ | |
| function Readability(doc, options) { | |
| // In some older versions, people passed a URI as the first argument. Cope: | |
| if (options && options.documentElement) { | |
| doc = options; | |
| options = arguments[2]; | |
| } else if (!doc || !doc.documentElement) { | |
| throw new Error( | |
| "First argument to Readability constructor should be a document object.", | |
| ); | |
| } | |
| options = options || {}; | |
| this._doc = doc; | |
| this._docJSDOMParser = this._doc.firstChild.__JSDOMParser__; | |
| this._articleTitle = null; | |
| this._articleByline = null; | |
| this._articleDir = null; | |
| this._articleSiteName = null; | |
| this._attempts = []; | |
| // Configurable options | |
| this._debug = !!options.debug; | |
| this._maxElemsToParse = | |
| options.maxElemsToParse || this.DEFAULT_MAX_ELEMS_TO_PARSE; | |
| this._nbTopCandidates = | |
| options.nbTopCandidates || this.DEFAULT_N_TOP_CANDIDATES; | |
| this._charThreshold = options.charThreshold || this.DEFAULT_CHAR_THRESHOLD; | |
| this._classesToPreserve = this.CLASSES_TO_PRESERVE.concat( | |
| options.classesToPreserve || [], | |
| ); | |
| this._keepClasses = !!options.keepClasses; | |
| this._serializer = | |
| options.serializer || | |
| function (el) { | |
| return el.innerHTML; | |
| }; | |
| this._disableJSONLD = !!options.disableJSONLD; | |
| this._allowedVideoRegex = options.allowedVideoRegex || this.REGEXPS.videos; | |
| // Start with all flags set | |
| this._flags = | |
| this.FLAG_STRIP_UNLIKELYS | | |
| this.FLAG_WEIGHT_CLASSES | | |
| this.FLAG_CLEAN_CONDITIONALLY; | |
| // Control whether log messages are sent to the console | |
| if (this._debug) { | |
| let logNode = function (node) { | |
| if (node.nodeType == node.TEXT_NODE) { | |
| return `${node.nodeName} ("${node.textContent}")`; | |
| } | |
| let attrPairs = Array.from(node.attributes || [], function (attr) { | |
| return `${attr.name}="${attr.value}"`; | |
| }).join(" "); | |
| return `<${node.localName} ${attrPairs}>`; | |
| }; | |
| this.log = function () { | |
| if (typeof console !== "undefined") { | |
| let args = Array.from(arguments, (arg) => { | |
| if (arg && arg.nodeType == this.ELEMENT_NODE) { | |
| return logNode(arg); | |
| } | |
| return arg; | |
| }); | |
| args.unshift("Reader: (Readability)"); | |
| console.log.apply(console, args); | |
| } else if (typeof dump !== "undefined") { | |
| /* global dump */ | |
| var msg = Array.prototype.map | |
| .call(arguments, function (x) { | |
| return x && x.nodeName ? logNode(x) : x; | |
| }) | |
| .join(" "); | |
| dump("Reader: (Readability) " + msg + "\n"); | |
| } | |
| }; | |
| } else { | |
| this.log = function () {}; | |
| } | |
| } | |
| Readability.prototype = { | |
| FLAG_STRIP_UNLIKELYS: 0x1, | |
| FLAG_WEIGHT_CLASSES: 0x2, | |
| FLAG_CLEAN_CONDITIONALLY: 0x4, | |
| // https://developer.mozilla.org/en-US/docs/Web/API/Node/nodeType | |
| ELEMENT_NODE: 1, | |
| TEXT_NODE: 3, | |
| // Max number of nodes supported by this parser. Default: 0 (no limit) | |
| DEFAULT_MAX_ELEMS_TO_PARSE: 0, | |
| // The number of top candidates to consider when analysing how | |
| // tight the competition is among candidates. | |
| DEFAULT_N_TOP_CANDIDATES: 5, | |
| // Element tags to score by default. | |
| DEFAULT_TAGS_TO_SCORE: "section,h2,h3,h4,h5,h6,p,td,pre" | |
| .toUpperCase() | |
| .split(","), | |
| // The default number of chars an article must have in order to return a result | |
| DEFAULT_CHAR_THRESHOLD: 500, | |
| // All of the regular expressions in use within readability. | |
| // Defined up here so we don't instantiate them repeatedly in loops. | |
| REGEXPS: { | |
| // NOTE: These two regular expressions are duplicated in | |
| // Readability-readerable.js. Please keep both copies in sync. | |
| unlikelyCandidates: | |
| /-ad-|ai2html|banner|breadcrumbs|combx|comment|community|cover-wrap|disqus|extra|footer|gdpr|header|legends|menu|related|remark|replies|rss|shoutbox|sidebar|skyscraper|social|sponsor|supplemental|ad-break|agegate|pagination|pager|popup|yom-remote/i, | |
| okMaybeItsACandidate: /and|article|body|column|content|main|shadow/i, | |
| positive: | |
| /article|body|content|entry|hentry|h-entry|main|page|pagination|post|text|blog|story/i, | |
| negative: | |
| /-ad-|hidden|^hid$| hid$| hid |^hid |banner|combx|comment|com-|contact|foot|footer|footnote|gdpr|masthead|media|meta|outbrain|promo|related|scroll|share|shoutbox|sidebar|skyscraper|sponsor|shopping|tags|tool|widget/i, | |
| extraneous: | |
| /print|archive|comment|discuss|e[\-]?mail|share|reply|all|login|sign|single|utility/i, | |
| byline: /byline|author|dateline|writtenby|p-author/i, | |
| replaceFonts: /<(\/?)font[^>]*>/gi, | |
| normalize: /\s{2,}/g, | |
| videos: | |
| /\/\/(www\.)?((dailymotion|youtube|youtube-nocookie|player\.vimeo|v\.qq)\.com|(archive|upload\.wikimedia)\.org|player\.twitch\.tv)/i, | |
| shareElements: /(\b|_)(share|sharedaddy)(\b|_)/i, | |
| nextLink: /(next|weiter|continue|>([^\|]|$)|»([^\|]|$))/i, | |
| prevLink: /(prev|earl|old|new|<|«)/i, | |
| tokenize: /\W+/g, | |
| whitespace: /^\s*$/, | |
| hasContent: /\S$/, | |
| hashUrl: /^#.+/, | |
| srcsetUrl: /(\S+)(\s+[\d.]+[xw])?(\s*(?:,|$))/g, | |
| b64DataUrl: /^data:\s*([^\s;,]+)\s*;\s*base64\s*,/i, | |
| // See: https://schema.org/Article | |
| jsonLdArticleTypes: | |
| /^Article|AdvertiserContentArticle|NewsArticle|AnalysisNewsArticle|AskPublicNewsArticle|BackgroundNewsArticle|OpinionNewsArticle|ReportageNewsArticle|ReviewNewsArticle|Report|SatiricalArticle|ScholarlyArticle|MedicalScholarlyArticle|SocialMediaPosting|BlogPosting|LiveBlogPosting|DiscussionForumPosting|TechArticle|APIReference$/, | |
| }, | |
| UNLIKELY_ROLES: [ | |
| "menu", | |
| "menubar", | |
| "complementary", | |
| "navigation", | |
| "alert", | |
| "alertdialog", | |
| "dialog", | |
| ], | |
| DIV_TO_P_ELEMS: new Set([ | |
| "BLOCKQUOTE", | |
| "DL", | |
| "DIV", | |
| "IMG", | |
| "OL", | |
| "P", | |
| "PRE", | |
| "TABLE", | |
| "UL", | |
| ]), | |
| ALTER_TO_DIV_EXCEPTIONS: ["DIV", "ARTICLE", "SECTION", "P"], | |
| PRESENTATIONAL_ATTRIBUTES: [ | |
| "align", | |
| "background", | |
| "bgcolor", | |
| "border", | |
| "cellpadding", | |
| "cellspacing", | |
| "frame", | |
| "hspace", | |
| "rules", | |
| "style", | |
| "valign", | |
| "vspace", | |
| ], | |
| DEPRECATED_SIZE_ATTRIBUTE_ELEMS: ["TABLE", "TH", "TD", "HR", "PRE"], | |
| // The commented out elements qualify as phrasing content but tend to be | |
| // removed by readability when put into paragraphs, so we ignore them here. | |
| PHRASING_ELEMS: [ | |
| // "CANVAS", "IFRAME", "SVG", "VIDEO", | |
| "ABBR", | |
| "AUDIO", | |
| "B", | |
| "BDO", | |
| "BR", | |
| "BUTTON", | |
| "CITE", | |
| "CODE", | |
| "DATA", | |
| "DATALIST", | |
| "DFN", | |
| "EM", | |
| "EMBED", | |
| "I", | |
| "IMG", | |
| "INPUT", | |
| "KBD", | |
| "LABEL", | |
| "MARK", | |
| "MATH", | |
| "METER", | |
| "NOSCRIPT", | |
| "OBJECT", | |
| "OUTPUT", | |
| "PROGRESS", | |
| "Q", | |
| "RUBY", | |
| "SAMP", | |
| "SCRIPT", | |
| "SELECT", | |
| "SMALL", | |
| "SPAN", | |
| "STRONG", | |
| "SUB", | |
| "SUP", | |
| "TEXTAREA", | |
| "TIME", | |
| "VAR", | |
| "WBR", | |
| ], | |
| // These are the classes that readability sets itself. | |
| CLASSES_TO_PRESERVE: ["page"], | |
| // These are the list of HTML entities that need to be escaped. | |
| HTML_ESCAPE_MAP: { | |
| lt: "<", | |
| gt: ">", | |
| amp: "&", | |
| quot: '"', | |
| apos: "'", | |
| }, | |
| /** | |
| * Run any post-process modifications to article content as necessary. | |
| * | |
| * @param Element | |
| * @return void | |
| **/ | |
| _postProcessContent: function (articleContent) { | |
| // Readability cannot open relative uris so we convert them to absolute uris. | |
| this._fixRelativeUris(articleContent); | |
| this._simplifyNestedElements(articleContent); | |
| if (!this._keepClasses) { | |
| // Remove classes. | |
| this._cleanClasses(articleContent); | |
| } | |
| }, | |
| /** | |
| * Iterates over a NodeList, calls `filterFn` for each node and removes node | |
| * if function returned `true`. | |
| * | |
| * If function is not passed, removes all the nodes in node list. | |
| * | |
| * @param NodeList nodeList The nodes to operate on | |
| * @param Function filterFn the function to use as a filter | |
| * @return void | |
| */ | |
| _removeNodes: function (nodeList, filterFn) { | |
| // Avoid ever operating on live node lists. | |
| if (this._docJSDOMParser && nodeList._isLiveNodeList) { | |
| throw new Error("Do not pass live node lists to _removeNodes"); | |
| } | |
| for (var i = nodeList.length - 1; i >= 0; i--) { | |
| var node = nodeList[i]; | |
| var parentNode = node.parentNode; | |
| if (parentNode) { | |
| if (!filterFn || filterFn.call(this, node, i, nodeList)) { | |
| parentNode.removeChild(node); | |
| } | |
| } | |
| } | |
| }, | |
| /** | |
| * Iterates over a NodeList, and calls _setNodeTag for each node. | |
| * | |
| * @param NodeList nodeList The nodes to operate on | |
| * @param String newTagName the new tag name to use | |
| * @return void | |
| */ | |
| _replaceNodeTags: function (nodeList, newTagName) { | |
| // Avoid ever operating on live node lists. | |
| if (this._docJSDOMParser && nodeList._isLiveNodeList) { | |
| throw new Error("Do not pass live node lists to _replaceNodeTags"); | |
| } | |
| for (const node of nodeList) { | |
| this._setNodeTag(node, newTagName); | |
| } | |
| }, | |
| /** | |
| * Iterate over a NodeList, which doesn't natively fully implement the Array | |
| * interface. | |
| * | |
| * For convenience, the current object context is applied to the provided | |
| * iterate function. | |
| * | |
| * @param NodeList nodeList The NodeList. | |
| * @param Function fn The iterate function. | |
| * @return void | |
| */ | |
| _forEachNode: function (nodeList, fn) { | |
| Array.prototype.forEach.call(nodeList, fn, this); | |
| }, | |
| /** | |
| * Iterate over a NodeList, and return the first node that passes | |
| * the supplied test function | |
| * | |
| * For convenience, the current object context is applied to the provided | |
| * test function. | |
| * | |
| * @param NodeList nodeList The NodeList. | |
| * @param Function fn The test function. | |
| * @return void | |
| */ | |
| _findNode: function (nodeList, fn) { | |
| return Array.prototype.find.call(nodeList, fn, this); | |
| }, | |
| /** | |
| * Iterate over a NodeList, return true if any of the provided iterate | |
| * function calls returns true, false otherwise. | |
| * | |
| * For convenience, the current object context is applied to the | |
| * provided iterate function. | |
| * | |
| * @param NodeList nodeList The NodeList. | |
| * @param Function fn The iterate function. | |
| * @return Boolean | |
| */ | |
| _someNode: function (nodeList, fn) { | |
| return Array.prototype.some.call(nodeList, fn, this); | |
| }, | |
| /** | |
| * Iterate over a NodeList, return true if all of the provided iterate | |
| * function calls return true, false otherwise. | |
| * | |
| * For convenience, the current object context is applied to the | |
| * provided iterate function. | |
| * | |
| * @param NodeList nodeList The NodeList. | |
| * @param Function fn The iterate function. | |
| * @return Boolean | |
| */ | |
| _everyNode: function (nodeList, fn) { | |
| return Array.prototype.every.call(nodeList, fn, this); | |
| }, | |
| /** | |
| * Concat all nodelists passed as arguments. | |
| * | |
| * @return ...NodeList | |
| * @return Array | |
| */ | |
| _concatNodeLists: function () { | |
| var slice = Array.prototype.slice; | |
| var args = slice.call(arguments); | |
| var nodeLists = args.map(function (list) { | |
| return slice.call(list); | |
| }); | |
| return Array.prototype.concat.apply([], nodeLists); | |
| }, | |
| _getAllNodesWithTag: function (node, tagNames) { | |
| if (node.querySelectorAll) { | |
| return node.querySelectorAll(tagNames.join(",")); | |
| } | |
| return [].concat.apply( | |
| [], | |
| tagNames.map(function (tag) { | |
| var collection = node.getElementsByTagName(tag); | |
| return Array.isArray(collection) ? collection : Array.from(collection); | |
| }), | |
| ); | |
| }, | |
| /** | |
| * Removes the class="" attribute from every element in the given | |
| * subtree, except those that match CLASSES_TO_PRESERVE and | |
| * the classesToPreserve array from the options object. | |
| * | |
| * @param Element | |
| * @return void | |
| */ | |
| _cleanClasses: function (node) { | |
| var classesToPreserve = this._classesToPreserve; | |
| var className = (node.getAttribute("class") || "") | |
| .split(/\s+/) | |
| .filter(function (cls) { | |
| return classesToPreserve.indexOf(cls) != -1; | |
| }) | |
| .join(" "); | |
| if (className) { | |
| node.setAttribute("class", className); | |
| } else { | |
| node.removeAttribute("class"); | |
| } | |
| for (node = node.firstElementChild; node; node = node.nextElementSibling) { | |
| this._cleanClasses(node); | |
| } | |
| }, | |
| /** | |
| * Converts each <a> and <img> uri in the given element to an absolute URI, | |
| * ignoring #ref URIs. | |
| * | |
| * @param Element | |
| * @return void | |
| */ | |
| _fixRelativeUris: function (articleContent) { | |
| var baseURI = this._doc.baseURI; | |
| var documentURI = this._doc.documentURI; | |
| function toAbsoluteURI(uri) { | |
| // Leave hash links alone if the base URI matches the document URI: | |
| if (baseURI == documentURI && uri.charAt(0) == "#") { | |
| return uri; | |
| } | |
| // Otherwise, resolve against base URI: | |
| try { | |
| return new URL(uri, baseURI).href; | |
| } catch (ex) { | |
| // Something went wrong, just return the original: | |
| } | |
| return uri; | |
| } | |
| var links = this._getAllNodesWithTag(articleContent, ["a"]); | |
| this._forEachNode(links, function (link) { | |
| var href = link.getAttribute("href"); | |
| if (href) { | |
| // Remove links with javascript: URIs, since | |
| // they won't work after scripts have been removed from the page. | |
| if (href.indexOf("javascript:") === 0) { | |
| // if the link only contains simple text content, it can be converted to a text node | |
| if ( | |
| link.childNodes.length === 1 && | |
| link.childNodes[0].nodeType === this.TEXT_NODE | |
| ) { | |
| var text = this._doc.createTextNode(link.textContent); | |
| link.parentNode.replaceChild(text, link); | |
| } else { | |
| // if the link has multiple children, they should all be preserved | |
| var container = this._doc.createElement("span"); | |
| while (link.firstChild) { | |
| container.appendChild(link.firstChild); | |
| } | |
| link.parentNode.replaceChild(container, link); | |
| } | |
| } else { | |
| link.setAttribute("href", toAbsoluteURI(href)); | |
| } | |
| } | |
| }); | |
| var mediaTypes = this._getAllNodesWithTag(articleContent, [ | |
| "img", | |
| "picture", | |
| "figure", | |
| "video", | |
| "audio", | |
| "source", | |
| ]); | |
| this._forEachNode(mediaTypes, function (media) { | |
| var src = media.getAttribute("src"); | |
| var poster = media.getAttribute("poster"); | |
| var srcset = media.getAttribute("srcset"); | |
| if (src) { | |
| media.setAttribute("src", toAbsoluteURI(src)); | |
| } | |
| if (poster) { | |
| media.setAttribute("poster", toAbsoluteURI(poster)); | |
| } | |
| if (srcset) { | |
| var newSrcset = srcset.replace( | |
| this.REGEXPS.srcsetUrl, | |
| function (_, p1, p2, p3) { | |
| return toAbsoluteURI(p1) + (p2 || "") + p3; | |
| }, | |
| ); | |
| media.setAttribute("srcset", newSrcset); | |
| } | |
| }); | |
| }, | |
| _simplifyNestedElements: function (articleContent) { | |
| var node = articleContent; | |
| while (node) { | |
| if ( | |
| node.parentNode && | |
| ["DIV", "SECTION"].includes(node.tagName) && | |
| !(node.id && node.id.startsWith("readability")) | |
| ) { | |
| if (this._isElementWithoutContent(node)) { | |
| node = this._removeAndGetNext(node); | |
| continue; | |
| } else if ( | |
| this._hasSingleTagInsideElement(node, "DIV") || | |
| this._hasSingleTagInsideElement(node, "SECTION") | |
| ) { | |
| var child = node.children[0]; | |
| for (var i = 0; i < node.attributes.length; i++) { | |
| child.setAttribute( | |
| node.attributes[i].name, | |
| node.attributes[i].value, | |
| ); | |
| } | |
| node.parentNode.replaceChild(child, node); | |
| node = child; | |
| continue; | |
| } | |
| } | |
| node = this._getNextNode(node); | |
| } | |
| }, | |
| /** | |
| * Get the article title as an H1. | |
| * | |
| * @return string | |
| **/ | |
| _getArticleTitle: function () { | |
| var doc = this._doc; | |
| var curTitle = ""; | |
| var origTitle = ""; | |
| try { | |
| curTitle = origTitle = doc.title.trim(); | |
| // If they had an element with id "title" in their HTML | |
| if (typeof curTitle !== "string") | |
| curTitle = origTitle = this._getInnerText( | |
| doc.getElementsByTagName("title")[0], | |
| ); | |
| } catch (e) { | |
| /* ignore exceptions setting the title. */ | |
| } | |
| var titleHadHierarchicalSeparators = false; | |
| function wordCount(str) { | |
| return str.split(/\s+/).length; | |
| } | |
| // If there's a separator in the title, first remove the final part | |
| if (/ [\|\-\\\/>»] /.test(curTitle)) { | |
| titleHadHierarchicalSeparators = / [\\\/>»] /.test(curTitle); | |
| curTitle = origTitle.replace(/(.*)[\|\-\\\/>»] .*/gi, "$1"); | |
| // If the resulting title is too short (3 words or fewer), remove | |
| // the first part instead: | |
| if (wordCount(curTitle) < 3) | |
| curTitle = origTitle.replace(/[^\|\-\\\/>»]*[\|\-\\\/>»](.*)/gi, "$1"); | |
| } else if (curTitle.indexOf(": ") !== -1) { | |
| // Check if we have an heading containing this exact string, so we | |
| // could assume it's the full title. | |
| var headings = this._concatNodeLists( | |
| doc.getElementsByTagName("h1"), | |
| doc.getElementsByTagName("h2"), | |
| ); | |
| var trimmedTitle = curTitle.trim(); | |
| var match = this._someNode(headings, function (heading) { | |
| return heading.textContent.trim() === trimmedTitle; | |
| }); | |
| // If we don't, let's extract the title out of the original title string. | |
| if (!match) { | |
| curTitle = origTitle.substring(origTitle.lastIndexOf(":") + 1); | |
| // If the title is now too short, try the first colon instead: | |
| if (wordCount(curTitle) < 3) { | |
| curTitle = origTitle.substring(origTitle.indexOf(":") + 1); | |
| // But if we have too many words before the colon there's something weird | |
| // with the titles and the H tags so let's just use the original title instead | |
| } else if (wordCount(origTitle.substr(0, origTitle.indexOf(":"))) > 5) { | |
| curTitle = origTitle; | |
| } | |
| } | |
| } else if (curTitle.length > 150 || curTitle.length < 15) { | |
| var hOnes = doc.getElementsByTagName("h1"); | |
| if (hOnes.length === 1) curTitle = this._getInnerText(hOnes[0]); | |
| } | |
| curTitle = curTitle.trim().replace(this.REGEXPS.normalize, " "); | |
| // If we now have 4 words or fewer as our title, and either no | |
| // 'hierarchical' separators (\, /, > or ») were found in the original | |
| // title or we decreased the number of words by more than 1 word, use | |
| // the original title. | |
| var curTitleWordCount = wordCount(curTitle); | |
| if ( | |
| curTitleWordCount <= 4 && | |
| (!titleHadHierarchicalSeparators || | |
| curTitleWordCount != | |
| wordCount(origTitle.replace(/[\|\-\\\/>»]+/g, "")) - 1) | |
| ) { | |
| curTitle = origTitle; | |
| } | |
| return curTitle; | |
| }, | |
| /** | |
| * Prepare the HTML document for readability to scrape it. | |
| * This includes things like stripping javascript, CSS, and handling terrible markup. | |
| * | |
| * @return void | |
| **/ | |
| _prepDocument: function () { | |
| var doc = this._doc; | |
| // Remove all style tags in head | |
| this._removeNodes(this._getAllNodesWithTag(doc, ["style"])); | |
| if (doc.body) { | |
| this._replaceBrs(doc.body); | |
| } | |
| this._replaceNodeTags(this._getAllNodesWithTag(doc, ["font"]), "SPAN"); | |
| }, | |
| /** | |
| * Finds the next node, starting from the given node, and ignoring | |
| * whitespace in between. If the given node is an element, the same node is | |
| * returned. | |
| */ | |
| _nextNode: function (node) { | |
| var next = node; | |
| while ( | |
| next && | |
| next.nodeType != this.ELEMENT_NODE && | |
| this.REGEXPS.whitespace.test(next.textContent) | |
| ) { | |
| next = next.nextSibling; | |
| } | |
| return next; | |
| }, | |
| /** | |
| * Replaces 2 or more successive <br> elements with a single <p>. | |
| * Whitespace between <br> elements are ignored. For example: | |
| * <div>foo<br>bar<br> <br><br>abc</div> | |
| * will become: | |
| * <div>foo<br>bar<p>abc</p></div> | |
| */ | |
| _replaceBrs: function (elem) { | |
| this._forEachNode(this._getAllNodesWithTag(elem, ["br"]), function (br) { | |
| var next = br.nextSibling; | |
| // Whether 2 or more <br> elements have been found and replaced with a | |
| // <p> block. | |
| var replaced = false; | |
| // If we find a <br> chain, remove the <br>s until we hit another node | |
| // or non-whitespace. This leaves behind the first <br> in the chain | |
| // (which will be replaced with a <p> later). | |
| while ((next = this._nextNode(next)) && next.tagName == "BR") { | |
| replaced = true; | |
| var brSibling = next.nextSibling; | |
| next.parentNode.removeChild(next); | |
| next = brSibling; | |
| } | |
| // If we removed a <br> chain, replace the remaining <br> with a <p>. Add | |
| // all sibling nodes as children of the <p> until we hit another <br> | |
| // chain. | |
| if (replaced) { | |
| var p = this._doc.createElement("p"); | |
| br.parentNode.replaceChild(p, br); | |
| next = p.nextSibling; | |
| while (next) { | |
| // If we've hit another <br><br>, we're done adding children to this <p>. | |
| if (next.tagName == "BR") { | |
| var nextElem = this._nextNode(next.nextSibling); | |
| if (nextElem && nextElem.tagName == "BR") break; | |
| } | |
| if (!this._isPhrasingContent(next)) break; | |
| // Otherwise, make this node a child of the new <p>. | |
| var sibling = next.nextSibling; | |
| p.appendChild(next); | |
| next = sibling; | |
| } | |
| while (p.lastChild && this._isWhitespace(p.lastChild)) { | |
| p.removeChild(p.lastChild); | |
| } | |
| if (p.parentNode.tagName === "P") this._setNodeTag(p.parentNode, "DIV"); | |
| } | |
| }); | |
| }, | |
| _setNodeTag: function (node, tag) { | |
| this.log("_setNodeTag", node, tag); | |
| if (this._docJSDOMParser) { | |
| node.localName = tag.toLowerCase(); | |
| node.tagName = tag.toUpperCase(); | |
| return node; | |
| } | |
| var replacement = node.ownerDocument.createElement(tag); | |
| while (node.firstChild) { | |
| replacement.appendChild(node.firstChild); | |
| } | |
| node.parentNode.replaceChild(replacement, node); | |
| if (node.readability) replacement.readability = node.readability; | |
| for (var i = 0; i < node.attributes.length; i++) { | |
| try { | |
| replacement.setAttribute( | |
| node.attributes[i].name, | |
| node.attributes[i].value, | |
| ); | |
| } catch (ex) { | |
| /* it's possible for setAttribute() to throw if the attribute name | |
| * isn't a valid XML Name. Such attributes can however be parsed from | |
| * source in HTML docs, see https://github.com/whatwg/html/issues/4275, | |
| * so we can hit them here and then throw. We don't care about such | |
| * attributes so we ignore them. | |
| */ | |
| } | |
| } | |
| return replacement; | |
| }, | |
| /** | |
| * Prepare the article node for display. Clean out any inline styles, | |
| * iframes, forms, strip extraneous <p> tags, etc. | |
| * | |
| * @param Element | |
| * @return void | |
| **/ | |
| _prepArticle: function (articleContent) { | |
| this._cleanStyles(articleContent); | |
| // Check for data tables before we continue, to avoid removing items in | |
| // those tables, which will often be isolated even though they're | |
| // visually linked to other content-ful elements (text, images, etc.). | |
| this._markDataTables(articleContent); | |
| this._fixLazyImages(articleContent); | |
| // Clean out junk from the article content | |
| this._cleanConditionally(articleContent, "form"); | |
| this._cleanConditionally(articleContent, "fieldset"); | |
| this._clean(articleContent, "object"); | |
| this._clean(articleContent, "embed"); | |
| this._clean(articleContent, "footer"); | |
| this._clean(articleContent, "link"); | |
| this._clean(articleContent, "aside"); | |
| // Clean out elements with little content that have "share" in their id/class combinations from final top candidates, | |
| // which means we don't remove the top candidates even they have "share". | |
| var shareElementThreshold = this.DEFAULT_CHAR_THRESHOLD; | |
| this._forEachNode(articleContent.children, function (topCandidate) { | |
| this._cleanMatchedNodes(topCandidate, function (node, matchString) { | |
| return ( | |
| this.REGEXPS.shareElements.test(matchString) && | |
| node.textContent.length < shareElementThreshold | |
| ); | |
| }); | |
| }); | |
| this._clean(articleContent, "iframe"); | |
| this._clean(articleContent, "input"); | |
| this._clean(articleContent, "textarea"); | |
| this._clean(articleContent, "select"); | |
| this._clean(articleContent, "button"); | |
| this._cleanHeaders(articleContent); | |
| // Do these last as the previous stuff may have removed junk | |
| // that will affect these | |
| this._cleanConditionally(articleContent, "table"); | |
| this._cleanConditionally(articleContent, "ul"); | |
| this._cleanConditionally(articleContent, "div"); | |
| // replace H1 with H2 as H1 should be only title that is displayed separately | |
| this._replaceNodeTags( | |
| this._getAllNodesWithTag(articleContent, ["h1"]), | |
| "h2", | |
| ); | |
| // Remove extra paragraphs | |
| this._removeNodes( | |
| this._getAllNodesWithTag(articleContent, ["p"]), | |
| function (paragraph) { | |
| var imgCount = paragraph.getElementsByTagName("img").length; | |
| var embedCount = paragraph.getElementsByTagName("embed").length; | |
| var objectCount = paragraph.getElementsByTagName("object").length; | |
| // At this point, nasty iframes have been removed, only remain embedded video ones. | |
| var iframeCount = paragraph.getElementsByTagName("iframe").length; | |
| var totalCount = imgCount + embedCount + objectCount + iframeCount; | |
| return totalCount === 0 && !this._getInnerText(paragraph, false); | |
| }, | |
| ); | |
| this._forEachNode( | |
| this._getAllNodesWithTag(articleContent, ["br"]), | |
| function (br) { | |
| var next = this._nextNode(br.nextSibling); | |
| if (next && next.tagName == "P") br.parentNode.removeChild(br); | |
| }, | |
| ); | |
| // Remove single-cell tables | |
| this._forEachNode( | |
| this._getAllNodesWithTag(articleContent, ["table"]), | |
| function (table) { | |
| var tbody = this._hasSingleTagInsideElement(table, "TBODY") | |
| ? table.firstElementChild | |
| : table; | |
| if (this._hasSingleTagInsideElement(tbody, "TR")) { | |
| var row = tbody.firstElementChild; | |
| if (this._hasSingleTagInsideElement(row, "TD")) { | |
| var cell = row.firstElementChild; | |
| cell = this._setNodeTag( | |
| cell, | |
| this._everyNode(cell.childNodes, this._isPhrasingContent) | |
| ? "P" | |
| : "DIV", | |
| ); | |
| table.parentNode.replaceChild(cell, table); | |
| } | |
| } | |
| }, | |
| ); | |
| }, | |
| /** | |
| * Initialize a node with the readability object. Also checks the | |
| * className/id for special names to add to its score. | |
| * | |
| * @param Element | |
| * @return void | |
| **/ | |
| _initializeNode: function (node) { | |
| node.readability = { contentScore: 0 }; | |
| switch (node.tagName) { | |
| case "DIV": | |
| node.readability.contentScore += 5; | |
| break; | |
| case "PRE": | |
| case "TD": | |
| case "BLOCKQUOTE": | |
| node.readability.contentScore += 3; | |
| break; | |
| case "ADDRESS": | |
| case "OL": | |
| case "UL": | |
| case "DL": | |
| case "DD": | |
| case "DT": | |
| case "LI": | |
| case "FORM": | |
| node.readability.contentScore -= 3; | |
| break; | |
| case "H1": | |
| case "H2": | |
| case "H3": | |
| case "H4": | |
| case "H5": | |
| case "H6": | |
| case "TH": | |
| node.readability.contentScore -= 5; | |
| break; | |
| } | |
| node.readability.contentScore += this._getClassWeight(node); | |
| }, | |
| _removeAndGetNext: function (node) { | |
| var nextNode = this._getNextNode(node, true); | |
| node.parentNode.removeChild(node); | |
| return nextNode; | |
| }, | |
| /** | |
| * Traverse the DOM from node to node, starting at the node passed in. | |
| * Pass true for the second parameter to indicate this node itself | |
| * (and its kids) are going away, and we want the next node over. | |
| * | |
| * Calling this in a loop will traverse the DOM depth-first. | |
| */ | |
| _getNextNode: function (node, ignoreSelfAndKids) { | |
| // First check for kids if those aren't being ignored | |
| if (!ignoreSelfAndKids && node.firstElementChild) { | |
| return node.firstElementChild; | |
| } | |
| // Then for siblings... | |
| if (node.nextElementSibling) { | |
| return node.nextElementSibling; | |
| } | |
| // And finally, move up the parent chain *and* find a sibling | |
| // (because this is depth-first traversal, we will have already | |
| // seen the parent nodes themselves). | |
| do { | |
| node = node.parentNode; | |
| } while (node && !node.nextElementSibling); | |
| return node && node.nextElementSibling; | |
| }, | |
| // compares second text to first one | |
| // 1 = same text, 0 = completely different text | |
| // works the way that it splits both texts into words and then finds words that are unique in second text | |
| // the result is given by the lower length of unique parts | |
| _textSimilarity: function (textA, textB) { | |
| var tokensA = textA | |
| .toLowerCase() | |
| .split(this.REGEXPS.tokenize) | |
| .filter(Boolean); | |
| var tokensB = textB | |
| .toLowerCase() | |
| .split(this.REGEXPS.tokenize) | |
| .filter(Boolean); | |
| if (!tokensA.length || !tokensB.length) { | |
| return 0; | |
| } | |
| var uniqTokensB = tokensB.filter((token) => !tokensA.includes(token)); | |
| var distanceB = uniqTokensB.join(" ").length / tokensB.join(" ").length; | |
| return 1 - distanceB; | |
| }, | |
| _checkByline: function (node, matchString) { | |
| if (this._articleByline) { | |
| return false; | |
| } | |
| if (node.getAttribute !== undefined) { | |
| var rel = node.getAttribute("rel"); | |
| var itemprop = node.getAttribute("itemprop"); | |
| } | |
| if ( | |
| (rel === "author" || | |
| (itemprop && itemprop.indexOf("author") !== -1) || | |
| this.REGEXPS.byline.test(matchString)) && | |
| this._isValidByline(node.textContent) | |
| ) { | |
| this._articleByline = node.textContent.trim(); | |
| return true; | |
| } | |
| return false; | |
| }, | |
| _getNodeAncestors: function (node, maxDepth) { | |
| maxDepth = maxDepth || 0; | |
| var i = 0, | |
| ancestors = []; | |
| while (node.parentNode) { | |
| ancestors.push(node.parentNode); | |
| if (maxDepth && ++i === maxDepth) break; | |
| node = node.parentNode; | |
| } | |
| return ancestors; | |
| }, | |
| /*** | |
| * grabArticle - Using a variety of metrics (content score, classname, element types), find the content that is | |
| * most likely to be the stuff a user wants to read. Then return it wrapped up in a div. | |
| * | |
| * @param page a document to run upon. Needs to be a full document, complete with body. | |
| * @return Element | |
| **/ | |
| _grabArticle: function (page) { | |
| this.log("**** grabArticle ****"); | |
| var doc = this._doc; | |
| var isPaging = page !== null; | |
| page = page ? page : this._doc.body; | |
| // We can't grab an article if we don't have a page! | |
| if (!page) { | |
| this.log("No body found in document. Abort."); | |
| return null; | |
| } | |
| var pageCacheHtml = page.innerHTML; | |
| while (true) { | |
| this.log("Starting grabArticle loop"); | |
| var stripUnlikelyCandidates = this._flagIsActive( | |
| this.FLAG_STRIP_UNLIKELYS, | |
| ); | |
| // First, node prepping. Trash nodes that look cruddy (like ones with the | |
| // class name "comment", etc), and turn divs into P tags where they have been | |
| // used inappropriately (as in, where they contain no other block level elements.) | |
| var elementsToScore = []; | |
| var node = this._doc.documentElement; | |
| let shouldRemoveTitleHeader = true; | |
| while (node) { | |
| if (node.tagName === "HTML") { | |
| this._articleLang = node.getAttribute("lang"); | |
| } | |
| var matchString = node.className + " " + node.id; | |
| if (!this._isProbablyVisible(node)) { | |
| this.log("Removing hidden node - " + matchString); | |
| node = this._removeAndGetNext(node); | |
| continue; | |
| } | |
| // User is not able to see elements applied with both "aria-modal = true" and "role = dialog" | |
| if ( | |
| node.getAttribute("aria-modal") == "true" && | |
| node.getAttribute("role") == "dialog" | |
| ) { | |
| node = this._removeAndGetNext(node); | |
| continue; | |
| } | |
| // Check to see if this node is a byline, and remove it if it is. | |
| if (this._checkByline(node, matchString)) { | |
| node = this._removeAndGetNext(node); | |
| continue; | |
| } | |
| if (shouldRemoveTitleHeader && this._headerDuplicatesTitle(node)) { | |
| this.log( | |
| "Removing header: ", | |
| node.textContent.trim(), | |
| this._articleTitle.trim(), | |
| ); | |
| shouldRemoveTitleHeader = false; | |
| node = this._removeAndGetNext(node); | |
| continue; | |
| } | |
| // Remove unlikely candidates | |
| if (stripUnlikelyCandidates) { | |
| if ( | |
| this.REGEXPS.unlikelyCandidates.test(matchString) && | |
| !this.REGEXPS.okMaybeItsACandidate.test(matchString) && | |
| !this._hasAncestorTag(node, "table") && | |
| !this._hasAncestorTag(node, "code") && | |
| node.tagName !== "BODY" && | |
| node.tagName !== "A" | |
| ) { | |
| this.log("Removing unlikely candidate - " + matchString); | |
| node = this._removeAndGetNext(node); | |
| continue; | |
| } | |
| if (this.UNLIKELY_ROLES.includes(node.getAttribute("role"))) { | |
| this.log( | |
| "Removing content with role " + | |
| node.getAttribute("role") + | |
| " - " + | |
| matchString, | |
| ); | |
| node = this._removeAndGetNext(node); | |
| continue; | |
| } | |
| } | |
| // Remove DIV, SECTION, and HEADER nodes without any content(e.g. text, image, video, or iframe). | |
| if ( | |
| (node.tagName === "DIV" || | |
| node.tagName === "SECTION" || | |
| node.tagName === "HEADER" || | |
| node.tagName === "H1" || | |
| node.tagName === "H2" || | |
| node.tagName === "H3" || | |
| node.tagName === "H4" || | |
| node.tagName === "H5" || | |
| node.tagName === "H6") && | |
| this._isElementWithoutContent(node) | |
| ) { | |
| node = this._removeAndGetNext(node); | |
| continue; | |
| } | |
| if (this.DEFAULT_TAGS_TO_SCORE.indexOf(node.tagName) !== -1) { | |
| elementsToScore.push(node); | |
| } | |
| // Turn all divs that don't have children block level elements into p's | |
| if (node.tagName === "DIV") { | |
| // Put phrasing content into paragraphs. | |
| var p = null; | |
| var childNode = node.firstChild; | |
| while (childNode) { | |
| var nextSibling = childNode.nextSibling; | |
| if (this._isPhrasingContent(childNode)) { | |
| if (p !== null) { | |
| p.appendChild(childNode); | |
| } else if (!this._isWhitespace(childNode)) { | |
| p = doc.createElement("p"); | |
| node.replaceChild(p, childNode); | |
| p.appendChild(childNode); | |
| } | |
| } else if (p !== null) { | |
| while (p.lastChild && this._isWhitespace(p.lastChild)) { | |
| p.removeChild(p.lastChild); | |
| } | |
| p = null; | |
| } | |
| childNode = nextSibling; | |
| } | |
| // Sites like http://mobile.slate.com encloses each paragraph with a DIV | |
| // element. DIVs with only a P element inside and no text content can be | |
| // safely converted into plain P elements to avoid confusing the scoring | |
| // algorithm with DIVs with are, in practice, paragraphs. | |
| if ( | |
| this._hasSingleTagInsideElement(node, "P") && | |
| this._getLinkDensity(node) < 0.25 | |
| ) { | |
| var newNode = node.children[0]; | |
| node.parentNode.replaceChild(newNode, node); | |
| node = newNode; | |
| elementsToScore.push(node); | |
| } else if (!this._hasChildBlockElement(node)) { | |
| node = this._setNodeTag(node, "P"); | |
| elementsToScore.push(node); | |
| } | |
| } | |
| node = this._getNextNode(node); | |
| } | |
| /** | |
| * Loop through all paragraphs, and assign a score to them based on how content-y they look. | |
| * Then add their score to their parent node. | |
| * | |
| * A score is determined by things like number of commas, class names, etc. Maybe eventually link density. | |
| **/ | |
| var candidates = []; | |
| this._forEachNode(elementsToScore, function (elementToScore) { | |
| if ( | |
| !elementToScore.parentNode || | |
| typeof elementToScore.parentNode.tagName === "undefined" | |
| ) | |
| return; | |
| // If this paragraph is less than 25 characters, don't even count it. | |
| var innerText = this._getInnerText(elementToScore); | |
| if (innerText.length < 25) return; | |
| // Exclude nodes with no ancestor. | |
| var ancestors = this._getNodeAncestors(elementToScore, 5); | |
| if (ancestors.length === 0) return; | |
| var contentScore = 0; | |
| // Add a point for the paragraph itself as a base. | |
| contentScore += 1; | |
| // Add points for any commas within this paragraph. | |
| contentScore += innerText.split(",").length; | |
| // For every 100 characters in this paragraph, add another point. Up to 3 points. | |
| contentScore += Math.min(Math.floor(innerText.length / 100), 3); | |
| // Initialize and score ancestors. | |
| this._forEachNode(ancestors, function (ancestor, level) { | |
| if ( | |
| !ancestor.tagName || | |
| !ancestor.parentNode || | |
| typeof ancestor.parentNode.tagName === "undefined" | |
| ) | |
| return; | |
| if (typeof ancestor.readability === "undefined") { | |
| this._initializeNode(ancestor); | |
| candidates.push(ancestor); | |
| } | |
| // Node score divider: | |
| // - parent: 1 (no division) | |
| // - grandparent: 2 | |
| // - great grandparent+: ancestor level * 3 | |
| if (level === 0) var scoreDivider = 1; | |
| else if (level === 1) scoreDivider = 2; | |
| else scoreDivider = level * 3; | |
| ancestor.readability.contentScore += contentScore / scoreDivider; | |
| }); | |
| }); | |
| // After we've calculated scores, loop through all of the possible | |
| // candidate nodes we found and find the one with the highest score. | |
| var topCandidates = []; | |
| for (var c = 0, cl = candidates.length; c < cl; c += 1) { | |
| var candidate = candidates[c]; | |
| // Scale the final candidates score based on link density. Good content | |
| // should have a relatively small link density (5% or less) and be mostly | |
| // unaffected by this operation. | |
| var candidateScore = | |
| candidate.readability.contentScore * | |
| (1 - this._getLinkDensity(candidate)); | |
| candidate.readability.contentScore = candidateScore; | |
| this.log("Candidate:", candidate, "with score " + candidateScore); | |
| for (var t = 0; t < this._nbTopCandidates; t++) { | |
| var aTopCandidate = topCandidates[t]; | |
| if ( | |
| !aTopCandidate || | |
| candidateScore > aTopCandidate.readability.contentScore | |
| ) { | |
| topCandidates.splice(t, 0, candidate); | |
| if (topCandidates.length > this._nbTopCandidates) | |
| topCandidates.pop(); | |
| break; | |
| } | |
| } | |
| } | |
| var topCandidate = topCandidates[0] || null; | |
| var neededToCreateTopCandidate = false; | |
| var parentOfTopCandidate; | |
| // If we still have no top candidate, just use the body as a last resort. | |
| // We also have to copy the body node so it is something we can modify. | |
| if (topCandidate === null || topCandidate.tagName === "BODY") { | |
| // Move all of the page's children into topCandidate | |
| topCandidate = doc.createElement("DIV"); | |
| neededToCreateTopCandidate = true; | |
| // Move everything (not just elements, also text nodes etc.) into the container | |
| // so we even include text directly in the body: | |
| while (page.firstChild) { | |
| this.log("Moving child out:", page.firstChild); | |
| topCandidate.appendChild(page.firstChild); | |
| } | |
| page.appendChild(topCandidate); | |
| this._initializeNode(topCandidate); | |
| } else if (topCandidate) { | |
| // Find a better top candidate node if it contains (at least three) nodes which belong to `topCandidates` array | |
| // and whose scores are quite closed with current `topCandidate` node. | |
| var alternativeCandidateAncestors = []; | |
| for (var i = 1; i < topCandidates.length; i++) { | |
| if ( | |
| topCandidates[i].readability.contentScore / | |
| topCandidate.readability.contentScore >= | |
| 0.75 | |
| ) { | |
| alternativeCandidateAncestors.push( | |
| this._getNodeAncestors(topCandidates[i]), | |
| ); | |
| } | |
| } | |
| var MINIMUM_TOPCANDIDATES = 3; | |
| if (alternativeCandidateAncestors.length >= MINIMUM_TOPCANDIDATES) { | |
| parentOfTopCandidate = topCandidate.parentNode; | |
| while (parentOfTopCandidate.tagName !== "BODY") { | |
| var listsContainingThisAncestor = 0; | |
| for ( | |
| var ancestorIndex = 0; | |
| ancestorIndex < alternativeCandidateAncestors.length && | |
| listsContainingThisAncestor < MINIMUM_TOPCANDIDATES; | |
| ancestorIndex++ | |
| ) { | |
| listsContainingThisAncestor += Number( | |
| alternativeCandidateAncestors[ancestorIndex].includes( | |
| parentOfTopCandidate, | |
| ), | |
| ); | |
| } | |
| if (listsContainingThisAncestor >= MINIMUM_TOPCANDIDATES) { | |
| topCandidate = parentOfTopCandidate; | |
| break; | |
| } | |
| parentOfTopCandidate = parentOfTopCandidate.parentNode; | |
| } | |
| } | |
| if (!topCandidate.readability) { | |
| this._initializeNode(topCandidate); | |
| } | |
| // Because of our bonus system, parents of candidates might have scores | |
| // themselves. They get half of the node. There won't be nodes with higher | |
| // scores than our topCandidate, but if we see the score going *up* in the first | |
| // few steps up the tree, that's a decent sign that there might be more content | |
| // lurking in other places that we want to unify in. The sibling stuff | |
| // below does some of that - but only if we've looked high enough up the DOM | |
| // tree. | |
| parentOfTopCandidate = topCandidate.parentNode; | |
| var lastScore = topCandidate.readability.contentScore; | |
| // The scores shouldn't get too low. | |
| var scoreThreshold = lastScore / 3; | |
| while (parentOfTopCandidate.tagName !== "BODY") { | |
| if (!parentOfTopCandidate.readability) { | |
| parentOfTopCandidate = parentOfTopCandidate.parentNode; | |
| continue; | |
| } | |
| var parentScore = parentOfTopCandidate.readability.contentScore; | |
| if (parentScore < scoreThreshold) break; | |
| if (parentScore > lastScore) { | |
| // Alright! We found a better parent to use. | |
| topCandidate = parentOfTopCandidate; | |
| break; | |
| } | |
| lastScore = parentOfTopCandidate.readability.contentScore; | |
| parentOfTopCandidate = parentOfTopCandidate.parentNode; | |
| } | |
| // If the top candidate is the only child, use parent instead. This will help sibling | |
| // joining logic when adjacent content is actually located in parent's sibling node. | |
| parentOfTopCandidate = topCandidate.parentNode; | |
| while ( | |
| parentOfTopCandidate.tagName != "BODY" && | |
| parentOfTopCandidate.children.length == 1 | |
| ) { | |
| topCandidate = parentOfTopCandidate; | |
| parentOfTopCandidate = topCandidate.parentNode; | |
| } | |
| if (!topCandidate.readability) { | |
| this._initializeNode(topCandidate); | |
| } | |
| } | |
| // Now that we have the top candidate, look through its siblings for content | |
| // that might also be related. Things like preambles, content split by ads | |
| // that we removed, etc. | |
| var articleContent = doc.createElement("DIV"); | |
| if (isPaging) articleContent.id = "readability-content"; | |
| var siblingScoreThreshold = Math.max( | |
| 10, | |
| topCandidate.readability.contentScore * 0.2, | |
| ); | |
| // Keep potential top candidate's parent node to try to get text direction of it later. | |
| parentOfTopCandidate = topCandidate.parentNode; | |
| var siblings = parentOfTopCandidate.children; | |
| for (var s = 0, sl = siblings.length; s < sl; s++) { | |
| var sibling = siblings[s]; | |
| var append = false; | |
| this.log( | |
| "Looking at sibling node:", | |
| sibling, | |
| sibling.readability | |
| ? "with score " + sibling.readability.contentScore | |
| : "", | |
| ); | |
| this.log( | |
| "Sibling has score", | |
| sibling.readability ? sibling.readability.contentScore : "Unknown", | |
| ); | |
| if (sibling === topCandidate) { | |
| append = true; | |
| } else { | |
| var contentBonus = 0; | |
| // Give a bonus if sibling nodes and top candidates have the example same classname | |
| if ( | |
| sibling.className === topCandidate.className && | |
| topCandidate.className !== "" | |
| ) | |
| contentBonus += topCandidate.readability.contentScore * 0.2; | |
| if ( | |
| sibling.readability && | |
| sibling.readability.contentScore + contentBonus >= | |
| siblingScoreThreshold | |
| ) { | |
| append = true; | |
| } else if (sibling.nodeName === "P") { | |
| var linkDensity = this._getLinkDensity(sibling); | |
| var nodeContent = this._getInnerText(sibling); | |
| var nodeLength = nodeContent.length; | |
| if (nodeLength > 80 && linkDensity < 0.25) { | |
| append = true; | |
| } else if ( | |
| nodeLength < 80 && | |
| nodeLength > 0 && | |
| linkDensity === 0 && | |
| nodeContent.search(/\.( |$)/) !== -1 | |
| ) { | |
| append = true; | |
| } | |
| } | |
| } | |
| if (append) { | |
| this.log("Appending node:", sibling); | |
| if (this.ALTER_TO_DIV_EXCEPTIONS.indexOf(sibling.nodeName) === -1) { | |
| // We have a node that isn't a common block level element, like a form or td tag. | |
| // Turn it into a div so it doesn't get filtered out later by accident. | |
| this.log("Altering sibling:", sibling, "to div."); | |
| sibling = this._setNodeTag(sibling, "DIV"); | |
| } | |
| articleContent.appendChild(sibling); | |
| // Fetch children again to make it compatible | |
| // with DOM parsers without live collection support. | |
| siblings = parentOfTopCandidate.children; | |
| // siblings is a reference to the children array, and | |
| // sibling is removed from the array when we call appendChild(). | |
| // As a result, we must revisit this index since the nodes | |
| // have been shifted. | |
| s -= 1; | |
| sl -= 1; | |
| } | |
| } | |
| if (this._debug) | |
| this.log("Article content pre-prep: " + articleContent.innerHTML); | |
| // So we have all of the content that we need. Now we clean it up for presentation. | |
| this._prepArticle(articleContent); | |
| if (this._debug) | |
| this.log("Article content post-prep: " + articleContent.innerHTML); | |
| if (neededToCreateTopCandidate) { | |
| // We already created a fake div thing, and there wouldn't have been any siblings left | |
| // for the previous loop, so there's no point trying to create a new div, and then | |
| // move all the children over. Just assign IDs and class names here. No need to append | |
| // because that already happened anyway. | |
| topCandidate.id = "readability-page-1"; | |
| topCandidate.className = "page"; | |
| } else { | |
| var div = doc.createElement("DIV"); | |
| div.id = "readability-page-1"; | |
| div.className = "page"; | |
| while (articleContent.firstChild) { | |
| div.appendChild(articleContent.firstChild); | |
| } | |
| articleContent.appendChild(div); | |
| } | |
| if (this._debug) | |
| this.log("Article content after paging: " + articleContent.innerHTML); | |
| var parseSuccessful = true; | |
| // Now that we've gone through the full algorithm, check to see if | |
| // we got any meaningful content. If we didn't, we may need to re-run | |
| // grabArticle with different flags set. This gives us a higher likelihood of | |
| // finding the content, and the sieve approach gives us a higher likelihood of | |
| // finding the -right- content. | |
| var textLength = this._getInnerText(articleContent, true).length; | |
| if (textLength < this._charThreshold) { | |
| parseSuccessful = false; | |
| page.innerHTML = pageCacheHtml; | |
| if (this._flagIsActive(this.FLAG_STRIP_UNLIKELYS)) { | |
| this._removeFlag(this.FLAG_STRIP_UNLIKELYS); | |
| this._attempts.push({ | |
| articleContent: articleContent, | |
| textLength: textLength, | |
| }); | |
| } else if (this._flagIsActive(this.FLAG_WEIGHT_CLASSES)) { | |
| this._removeFlag(this.FLAG_WEIGHT_CLASSES); | |
| this._attempts.push({ | |
| articleContent: articleContent, | |
| textLength: textLength, | |
| }); | |
| } else if (this._flagIsActive(this.FLAG_CLEAN_CONDITIONALLY)) { | |
| this._removeFlag(this.FLAG_CLEAN_CONDITIONALLY); | |
| this._attempts.push({ | |
| articleContent: articleContent, | |
| textLength: textLength, | |
| }); | |
| } else { | |
| this._attempts.push({ | |
| articleContent: articleContent, | |
| textLength: textLength, | |
| }); | |
| // No luck after removing flags, just return the longest text we found during the different loops | |
| this._attempts.sort(function (a, b) { | |
| return b.textLength - a.textLength; | |
| }); | |
| // But first check if we actually have something | |
| if (!this._attempts[0].textLength) { | |
| return null; | |
| } | |
| articleContent = this._attempts[0].articleContent; | |
| parseSuccessful = true; | |
| } | |
| } | |
| if (parseSuccessful) { | |
| // Find out text direction from ancestors of final top candidate. | |
| var ancestors = [parentOfTopCandidate, topCandidate].concat( | |
| this._getNodeAncestors(parentOfTopCandidate), | |
| ); | |
| this._someNode(ancestors, function (ancestor) { | |
| if (!ancestor.tagName) return false; | |
| var articleDir = ancestor.getAttribute("dir"); | |
| if (articleDir) { | |
| this._articleDir = articleDir; | |
| return true; | |
| } | |
| return false; | |
| }); | |
| return articleContent; | |
| } | |
| } | |
| }, | |
| /** | |
| * Check whether the input string could be a byline. | |
| * This verifies that the input is a string, and that the length | |
| * is less than 100 chars. | |
| * | |
| * @param possibleByline {string} - a string to check whether its a byline. | |
| * @return Boolean - whether the input string is a byline. | |
| */ | |
| _isValidByline: function (byline) { | |
| if (typeof byline == "string" || byline instanceof String) { | |
| byline = byline.trim(); | |
| return byline.length > 0 && byline.length < 100; | |
| } | |
| return false; | |
| }, | |
| /** | |
| * Converts some of the common HTML entities in string to their corresponding characters. | |
| * | |
| * @param str {string} - a string to unescape. | |
| * @return string without HTML entity. | |
| */ | |
| _unescapeHtmlEntities: function (str) { | |
| if (!str) { | |
| return str; | |
| } | |
| var htmlEscapeMap = this.HTML_ESCAPE_MAP; | |
| return str | |
| .replace(/&(quot|amp|apos|lt|gt);/g, function (_, tag) { | |
| return htmlEscapeMap[tag]; | |
| }) | |
| .replace( | |
| /&#(?:x([0-9a-z]{1,4})|([0-9]{1,4}));/gi, | |
| function (_, hex, numStr) { | |
| var num = parseInt(hex || numStr, hex ? 16 : 10); | |
| return String.fromCharCode(num); | |
| }, | |
| ); | |
| }, | |
| /** | |
| * Try to extract metadata from JSON-LD object. | |
| * For now, only Schema.org objects of type Article or its subtypes are supported. | |
| * @return Object with any metadata that could be extracted (possibly none) | |
| */ | |
| _getJSONLD: function (doc) { | |
| var scripts = this._getAllNodesWithTag(doc, ["script"]); | |
| var metadata; | |
| this._forEachNode(scripts, function (jsonLdElement) { | |
| if ( | |
| !metadata && | |
| jsonLdElement.getAttribute("type") === "application/ld+json" | |
| ) { | |
| try { | |
| // Strip CDATA markers if present | |
| var content = jsonLdElement.textContent.replace( | |
| /^\s*<!\[CDATA\[|\]\]>\s*$/g, | |
| "", | |
| ); | |
| var parsed = JSON.parse(content); | |
| if ( | |
| !parsed["@context"] || | |
| !parsed["@context"].match(/^https?\:\/\/schema\.org$/) | |
| ) { | |
| return; | |
| } | |
| if (!parsed["@type"] && Array.isArray(parsed["@graph"])) { | |
| parsed = parsed["@graph"].find(function (it) { | |
| return (it["@type"] || "").match(this.REGEXPS.jsonLdArticleTypes); | |
| }); | |
| } | |
| if ( | |
| !parsed || | |
| !parsed["@type"] || | |
| !parsed["@type"].match(this.REGEXPS.jsonLdArticleTypes) | |
| ) { | |
| return; | |
| } | |
| metadata = {}; | |
| if ( | |
| typeof parsed.name === "string" && | |
| typeof parsed.headline === "string" && | |
| parsed.name !== parsed.headline | |
| ) { | |
| // we have both name and headline element in the JSON-LD. They should both be the same but some websites like aktualne.cz | |
| // put their own name into "name" and the article title to "headline" which confuses Readability. So we try to check if either | |
| // "name" or "headline" closely matches the html title, and if so, use that one. If not, then we use "name" by default. | |
| var title = this._getArticleTitle(); | |
| var nameMatches = this._textSimilarity(parsed.name, title) > 0.75; | |
| var headlineMatches = | |
| this._textSimilarity(parsed.headline, title) > 0.75; | |
| if (headlineMatches && !nameMatches) { | |
| metadata.title = parsed.headline; | |
| } else { | |
| metadata.title = parsed.name; | |
| } | |
| } else if (typeof parsed.name === "string") { | |
| metadata.title = parsed.name.trim(); | |
| } else if (typeof parsed.headline === "string") { | |
| metadata.title = parsed.headline.trim(); | |
| } | |
| if (parsed.author) { | |
| if (typeof parsed.author.name === "string") { | |
| metadata.byline = parsed.author.name.trim(); | |
| } else if ( | |
| Array.isArray(parsed.author) && | |
| parsed.author[0] && | |
| typeof parsed.author[0].name === "string" | |
| ) { | |
| metadata.byline = parsed.author | |
| .filter(function (author) { | |
| return author && typeof author.name === "string"; | |
| }) | |
| .map(function (author) { | |
| return author.name.trim(); | |
| }) | |
| .join(", "); | |
| } | |
| } | |
| if (typeof parsed.description === "string") { | |
| metadata.excerpt = parsed.description.trim(); | |
| } | |
| if (parsed.publisher && typeof parsed.publisher.name === "string") { | |
| metadata.siteName = parsed.publisher.name.trim(); | |
| } | |
| return; | |
| } catch (err) { | |
| this.log(err.message); | |
| } | |
| } | |
| }); | |
| return metadata ? metadata : {}; | |
| }, | |
| /** | |
| * Attempts to get excerpt and byline metadata for the article. | |
| * | |
| * @param {Object} jsonld — object containing any metadata that | |
| * could be extracted from JSON-LD object. | |
| * | |
| * @return Object with optional "excerpt" and "byline" properties | |
| */ | |
| _getArticleMetadata: function (jsonld) { | |
| var metadata = {}; | |
| var values = {}; | |
| var metaElements = this._doc.getElementsByTagName("meta"); | |
| // property is a space-separated list of values | |
| var propertyPattern = | |
| /\s*(dc|dcterm|og|twitter)\s*:\s*(author|creator|description|title|site_name)\s*/gi; | |
| // name is a single value | |
| var namePattern = | |
| /^\s*(?:(dc|dcterm|og|twitter|weibo:(article|webpage))\s*[\.:]\s*)?(author|creator|description|title|site_name)\s*$/i; | |
| // Find description tags. | |
| this._forEachNode(metaElements, function (element) { | |
| var elementName = element.getAttribute("name"); | |
| var elementProperty = element.getAttribute("property"); | |
| var content = element.getAttribute("content"); | |
| if (!content) { | |
| return; | |
| } | |
| var matches = null; | |
| var name = null; | |
| if (elementProperty) { | |
| matches = elementProperty.match(propertyPattern); | |
| if (matches) { | |
| // Convert to lowercase, and remove any whitespace | |
| // so we can match below. | |
| name = matches[0].toLowerCase().replace(/\s/g, ""); | |
| // multiple authors | |
| values[name] = content.trim(); | |
| } | |
| } | |
| if (!matches && elementName && namePattern.test(elementName)) { | |
| name = elementName; | |
| if (content) { | |
| // Convert to lowercase, remove any whitespace, and convert dots | |
| // to colons so we can match below. | |
| name = name.toLowerCase().replace(/\s/g, "").replace(/\./g, ":"); | |
| values[name] = content.trim(); | |
| } | |
| } | |
| }); | |
| // get title | |
| metadata.title = | |
| jsonld.title || | |
| values["dc:title"] || | |
| values["dcterm:title"] || | |
| values["og:title"] || | |
| values["weibo:article:title"] || | |
| values["weibo:webpage:title"] || | |
| values["title"] || | |
| values["twitter:title"]; | |
| if (!metadata.title) { | |
| metadata.title = this._getArticleTitle(); | |
| } | |
| // get author | |
| metadata.byline = | |
| jsonld.byline || | |
| values["dc:creator"] || | |
| values["dcterm:creator"] || | |
| values["author"]; | |
| // get description | |
| metadata.excerpt = | |
| jsonld.excerpt || | |
| values["dc:description"] || | |
| values["dcterm:description"] || | |
| values["og:description"] || | |
| values["weibo:article:description"] || | |
| values["weibo:webpage:description"] || | |
| values["description"] || | |
| values["twitter:description"]; | |
| // get site name | |
| metadata.siteName = jsonld.siteName || values["og:site_name"]; | |
| // in many sites the meta value is escaped with HTML entities, | |
| // so here we need to unescape it | |
| metadata.title = this._unescapeHtmlEntities(metadata.title); | |
| metadata.byline = this._unescapeHtmlEntities(metadata.byline); | |
| metadata.excerpt = this._unescapeHtmlEntities(metadata.excerpt); | |
| metadata.siteName = this._unescapeHtmlEntities(metadata.siteName); | |
| return metadata; | |
| }, | |
| /** | |
| * Check if node is image, or if node contains exactly only one image | |
| * whether as a direct child or as its descendants. | |
| * | |
| * @param Element | |
| **/ | |
| _isSingleImage: function (node) { | |
| if (node.tagName === "IMG") { | |
| return true; | |
| } | |
| if (node.children.length !== 1 || node.textContent.trim() !== "") { | |
| return false; | |
| } | |
| return this._isSingleImage(node.children[0]); | |
| }, | |
| /** | |
| * Find all <noscript> that are located after <img> nodes, and which contain only one | |
| * <img> element. Replace the first image with the image from inside the <noscript> tag, | |
| * and remove the <noscript> tag. This improves the quality of the images we use on | |
| * some sites (e.g. Medium). | |
| * | |
| * @param Element | |
| **/ | |
| _unwrapNoscriptImages: function (doc) { | |
| // Find img without source or attributes that might contains image, and remove it. | |
| // This is done to prevent a placeholder img is replaced by img from noscript in next step. | |
| var imgs = Array.from(doc.getElementsByTagName("img")); | |
| this._forEachNode(imgs, function (img) { | |
| for (var i = 0; i < img.attributes.length; i++) { | |
| var attr = img.attributes[i]; | |
| switch (attr.name) { | |
| case "src": | |
| case "srcset": | |
| case "data-src": | |
| case "data-srcset": | |
| return; | |
| } | |
| if (/\.(jpg|jpeg|png|webp)/i.test(attr.value)) { | |
| return; | |
| } | |
| } | |
| img.parentNode.removeChild(img); | |
| }); | |
| // Next find noscript and try to extract its image | |
| var noscripts = Array.from(doc.getElementsByTagName("noscript")); | |
| this._forEachNode(noscripts, function (noscript) { | |
| // Parse content of noscript and make sure it only contains image | |
| var tmp = doc.createElement("div"); | |
| tmp.innerHTML = noscript.innerHTML; | |
| if (!this._isSingleImage(tmp)) { | |
| return; | |
| } | |
| // If noscript has previous sibling and it only contains image, | |
| // replace it with noscript content. However we also keep old | |
| // attributes that might contains image. | |
| var prevElement = noscript.previousElementSibling; | |
| if (prevElement && this._isSingleImage(prevElement)) { | |
| var prevImg = prevElement; | |
| if (prevImg.tagName !== "IMG") { | |
| prevImg = prevElement.getElementsByTagName("img")[0]; | |
| } | |
| var newImg = tmp.getElementsByTagName("img")[0]; | |
| for (var i = 0; i < prevImg.attributes.length; i++) { | |
| var attr = prevImg.attributes[i]; | |
| if (attr.value === "") { | |
| continue; | |
| } | |
| if ( | |
| attr.name === "src" || | |
| attr.name === "srcset" || | |
| /\.(jpg|jpeg|png|webp)/i.test(attr.value) | |
| ) { | |
| if (newImg.getAttribute(attr.name) === attr.value) { | |
| continue; | |
| } | |
| var attrName = attr.name; | |
| if (newImg.hasAttribute(attrName)) { | |
| attrName = "data-old-" + attrName; | |
| } | |
| newImg.setAttribute(attrName, attr.value); | |
| } | |
| } | |
| noscript.parentNode.replaceChild(tmp.firstElementChild, prevElement); | |
| } | |
| }); | |
| }, | |
| /** | |
| * Removes script tags from the document. | |
| * | |
| * @param Element | |
| **/ | |
| _removeScripts: function (doc) { | |
| this._removeNodes(this._getAllNodesWithTag(doc, ["script", "noscript"])); | |
| }, | |
| /** | |
| * Check if this node has only whitespace and a single element with given tag | |
| * Returns false if the DIV node contains non-empty text nodes | |
| * or if it contains no element with given tag or more than 1 element. | |
| * | |
| * @param Element | |
| * @param string tag of child element | |
| **/ | |
| _hasSingleTagInsideElement: function (element, tag) { | |
| // There should be exactly 1 element child with given tag | |
| if (element.children.length != 1 || element.children[0].tagName !== tag) { | |
| return false; | |
| } | |
| // And there should be no text nodes with real content | |
| return !this._someNode(element.childNodes, function (node) { | |
| return ( | |
| node.nodeType === this.TEXT_NODE && | |
| this.REGEXPS.hasContent.test(node.textContent) | |
| ); | |
| }); | |
| }, | |
| _isElementWithoutContent: function (node) { | |
| return ( | |
| node.nodeType === this.ELEMENT_NODE && | |
| node.textContent.trim().length == 0 && | |
| (node.children.length == 0 || | |
| node.children.length == | |
| node.getElementsByTagName("br").length + | |
| node.getElementsByTagName("hr").length) | |
| ); | |
| }, | |
| /** | |
| * Determine whether element has any children block level elements. | |
| * | |
| * @param Element | |
| */ | |
| _hasChildBlockElement: function (element) { | |
| return this._someNode(element.childNodes, function (node) { | |
| return ( | |
| this.DIV_TO_P_ELEMS.has(node.tagName) || | |
| this._hasChildBlockElement(node) | |
| ); | |
| }); | |
| }, | |
| /*** | |
| * Determine if a node qualifies as phrasing content. | |
| * https://developer.mozilla.org/en-US/docs/Web/Guide/HTML/Content_categories#Phrasing_content | |
| **/ | |
| _isPhrasingContent: function (node) { | |
| return ( | |
| node.nodeType === this.TEXT_NODE || | |
| this.PHRASING_ELEMS.indexOf(node.tagName) !== -1 || | |
| ((node.tagName === "A" || | |
| node.tagName === "DEL" || | |
| node.tagName === "INS") && | |
| this._everyNode(node.childNodes, this._isPhrasingContent)) | |
| ); | |
| }, | |
| _isWhitespace: function (node) { | |
| return ( | |
| (node.nodeType === this.TEXT_NODE && | |
| node.textContent.trim().length === 0) || | |
| (node.nodeType === this.ELEMENT_NODE && node.tagName === "BR") | |
| ); | |
| }, | |
| /** | |
| * Get the inner text of a node - cross browser compatibly. | |
| * This also strips out any excess whitespace to be found. | |
| * | |
| * @param Element | |
| * @param Boolean normalizeSpaces (default: true) | |
| * @return string | |
| **/ | |
| _getInnerText: function (e, normalizeSpaces) { | |
| normalizeSpaces = | |
| typeof normalizeSpaces === "undefined" ? true : normalizeSpaces; | |
| var textContent = e.textContent.trim(); | |
| if (normalizeSpaces) { | |
| return textContent.replace(this.REGEXPS.normalize, " "); | |
| } | |
| return textContent; | |
| }, | |
| /** | |
| * Get the number of times a string s appears in the node e. | |
| * | |
| * @param Element | |
| * @param string - what to split on. Default is "," | |
| * @return number (integer) | |
| **/ | |
| _getCharCount: function (e, s) { | |
| s = s || ","; | |
| return this._getInnerText(e).split(s).length - 1; | |
| }, | |
| /** | |
| * Remove the style attribute on every e and under. | |
| * TODO: Test if getElementsByTagName(*) is faster. | |
| * | |
| * @param Element | |
| * @return void | |
| **/ | |
| _cleanStyles: function (e) { | |
| if (!e || e.tagName.toLowerCase() === "svg") return; | |
| // Remove `style` and deprecated presentational attributes | |
| for (var i = 0; i < this.PRESENTATIONAL_ATTRIBUTES.length; i++) { | |
| e.removeAttribute(this.PRESENTATIONAL_ATTRIBUTES[i]); | |
| } | |
| if (this.DEPRECATED_SIZE_ATTRIBUTE_ELEMS.indexOf(e.tagName) !== -1) { | |
| e.removeAttribute("width"); | |
| e.removeAttribute("height"); | |
| } | |
| var cur = e.firstElementChild; | |
| while (cur !== null) { | |
| this._cleanStyles(cur); | |
| cur = cur.nextElementSibling; | |
| } | |
| }, | |
| /** | |
| * Get the density of links as a percentage of the content | |
| * This is the amount of text that is inside a link divided by the total text in the node. | |
| * | |
| * @param Element | |
| * @return number (float) | |
| **/ | |
| _getLinkDensity: function (element) { | |
| var textLength = this._getInnerText(element).length; | |
| if (textLength === 0) return 0; | |
| var linkLength = 0; | |
| // XXX implement _reduceNodeList? | |
| this._forEachNode(element.getElementsByTagName("a"), function (linkNode) { | |
| var href = linkNode.getAttribute("href"); | |
| var coefficient = href && this.REGEXPS.hashUrl.test(href) ? 0.3 : 1; | |
| linkLength += this._getInnerText(linkNode).length * coefficient; | |
| }); | |
| return linkLength / textLength; | |
| }, | |
| /** | |
| * Get an elements class/id weight. Uses regular expressions to tell if this | |
| * element looks good or bad. | |
| * | |
| * @param Element | |
| * @return number (Integer) | |
| **/ | |
| _getClassWeight: function (e) { | |
| if (!this._flagIsActive(this.FLAG_WEIGHT_CLASSES)) return 0; | |
| var weight = 0; | |
| // Look for a special classname | |
| if (typeof e.className === "string" && e.className !== "") { | |
| if (this.REGEXPS.negative.test(e.className)) weight -= 25; | |
| if (this.REGEXPS.positive.test(e.className)) weight += 25; | |
| } | |
| // Look for a special ID | |
| if (typeof e.id === "string" && e.id !== "") { | |
| if (this.REGEXPS.negative.test(e.id)) weight -= 25; | |
| if (this.REGEXPS.positive.test(e.id)) weight += 25; | |
| } | |
| return weight; | |
| }, | |
| /** | |
| * Clean a node of all elements of type "tag". | |
| * (Unless it's a youtube/vimeo video. People love movies.) | |
| * | |
| * @param Element | |
| * @param string tag to clean | |
| * @return void | |
| **/ | |
| _clean: function (e, tag) { | |
| var isEmbed = ["object", "embed", "iframe"].indexOf(tag) !== -1; | |
| this._removeNodes(this._getAllNodesWithTag(e, [tag]), function (element) { | |
| // Allow youtube and vimeo videos through as people usually want to see those. | |
| if (isEmbed) { | |
| // First, check the elements attributes to see if any of them contain youtube or vimeo | |
| for (var i = 0; i < element.attributes.length; i++) { | |
| if (this._allowedVideoRegex.test(element.attributes[i].value)) { | |
| return false; | |
| } | |
| } | |
| // For embed with <object> tag, check inner HTML as well. | |
| if ( | |
| element.tagName === "object" && | |
| this._allowedVideoRegex.test(element.innerHTML) | |
| ) { | |
| return false; | |
| } | |
| } | |
| return true; | |
| }); | |
| }, | |
| /** | |
| * Check if a given node has one of its ancestor tag name matching the | |
| * provided one. | |
| * @param HTMLElement node | |
| * @param String tagName | |
| * @param Number maxDepth | |
| * @param Function filterFn a filter to invoke to determine whether this node 'counts' | |
| * @return Boolean | |
| */ | |
| _hasAncestorTag: function (node, tagName, maxDepth, filterFn) { | |
| maxDepth = maxDepth || 3; | |
| tagName = tagName.toUpperCase(); | |
| var depth = 0; | |
| while (node.parentNode) { | |
| if (maxDepth > 0 && depth > maxDepth) return false; | |
| if ( | |
| node.parentNode.tagName === tagName && | |
| (!filterFn || filterFn(node.parentNode)) | |
| ) | |
| return true; | |
| node = node.parentNode; | |
| depth++; | |
| } | |
| return false; | |
| }, | |
| /** | |
| * Return an object indicating how many rows and columns this table has. | |
| */ | |
| _getRowAndColumnCount: function (table) { | |
| var rows = 0; | |
| var columns = 0; | |
| var trs = table.getElementsByTagName("tr"); | |
| for (var i = 0; i < trs.length; i++) { | |
| var rowspan = trs[i].getAttribute("rowspan") || 0; | |
| if (rowspan) { | |
| rowspan = parseInt(rowspan, 10); | |
| } | |
| rows += rowspan || 1; | |
| // Now look for column-related info | |
| var columnsInThisRow = 0; | |
| var cells = trs[i].getElementsByTagName("td"); | |
| for (var j = 0; j < cells.length; j++) { | |
| var colspan = cells[j].getAttribute("colspan") || 0; | |
| if (colspan) { | |
| colspan = parseInt(colspan, 10); | |
| } | |
| columnsInThisRow += colspan || 1; | |
| } | |
| columns = Math.max(columns, columnsInThisRow); | |
| } | |
| return { rows: rows, columns: columns }; | |
| }, | |
| /** | |
| * Look for 'data' (as opposed to 'layout') tables, for which we use | |
| * similar checks as | |
| * https://searchfox.org/mozilla-central/rev/f82d5c549f046cb64ce5602bfd894b7ae807c8f8/accessible/generic/TableAccessible.cpp#19 | |
| */ | |
| _markDataTables: function (root) { | |
| var tables = root.getElementsByTagName("table"); | |
| for (var i = 0; i < tables.length; i++) { | |
| var table = tables[i]; | |
| var role = table.getAttribute("role"); | |
| if (role == "presentation") { | |
| table._readabilityDataTable = false; | |
| continue; | |
| } | |
| var datatable = table.getAttribute("datatable"); | |
| if (datatable == "0") { | |
| table._readabilityDataTable = false; | |
| continue; | |
| } | |
| var summary = table.getAttribute("summary"); | |
| if (summary) { | |
| table._readabilityDataTable = true; | |
| continue; | |
| } | |
| var caption = table.getElementsByTagName("caption")[0]; | |
| if (caption && caption.childNodes.length > 0) { | |
| table._readabilityDataTable = true; | |
| continue; | |
| } | |
| // If the table has a descendant with any of these tags, consider a data table: | |
| var dataTableDescendants = ["col", "colgroup", "tfoot", "thead", "th"]; | |
| var descendantExists = function (tag) { | |
| return !!table.getElementsByTagName(tag)[0]; | |
| }; | |
| if (dataTableDescendants.some(descendantExists)) { | |
| this.log("Data table because found data-y descendant"); | |
| table._readabilityDataTable = true; | |
| continue; | |
| } | |
| // Nested tables indicate a layout table: | |
| if (table.getElementsByTagName("table")[0]) { | |
| table._readabilityDataTable = false; | |
| continue; | |
| } | |
| var sizeInfo = this._getRowAndColumnCount(table); | |
| if (sizeInfo.rows >= 10 || sizeInfo.columns > 4) { | |
| table._readabilityDataTable = true; | |
| continue; | |
| } | |
| // Now just go by size entirely: | |
| table._readabilityDataTable = sizeInfo.rows * sizeInfo.columns > 10; | |
| } | |
| }, | |
| /* convert images and figures that have properties like data-src into images that can be loaded without JS */ | |
| _fixLazyImages: function (root) { | |
| this._forEachNode( | |
| this._getAllNodesWithTag(root, ["img", "picture", "figure"]), | |
| function (elem) { | |
| // In some sites (e.g. Kotaku), they put 1px square image as base64 data uri in the src attribute. | |
| // So, here we check if the data uri is too short, just might as well remove it. | |
| if (elem.src && this.REGEXPS.b64DataUrl.test(elem.src)) { | |
| // Make sure it's not SVG, because SVG can have a meaningful image in under 133 bytes. | |
| var parts = this.REGEXPS.b64DataUrl.exec(elem.src); | |
| if (parts[1] === "image/svg+xml") { | |
| return; | |
| } | |
| // Make sure this element has other attributes which contains image. | |
| // If it doesn't, then this src is important and shouldn't be removed. | |
| var srcCouldBeRemoved = false; | |
| for (var i = 0; i < elem.attributes.length; i++) { | |
| var attr = elem.attributes[i]; | |
| if (attr.name === "src") { | |
| continue; | |
| } | |
| if (/\.(jpg|jpeg|png|webp)/i.test(attr.value)) { | |
| srcCouldBeRemoved = true; | |
| break; | |
| } | |
| } | |
| // Here we assume if image is less than 100 bytes (or 133B after encoded to base64) | |
| // it will be too small, therefore it might be placeholder image. | |
| if (srcCouldBeRemoved) { | |
| var b64starts = elem.src.search(/base64\s*/i) + 7; | |
| var b64length = elem.src.length - b64starts; | |
| if (b64length < 133) { | |
| elem.removeAttribute("src"); | |
| } | |
| } | |
| } | |
| // also check for "null" to work around https://github.com/jsdom/jsdom/issues/2580 | |
| if ( | |
| (elem.src || (elem.srcset && elem.srcset != "null")) && | |
| elem.className.toLowerCase().indexOf("lazy") === -1 | |
| ) { | |
| return; | |
| } | |
| for (var j = 0; j < elem.attributes.length; j++) { | |
| attr = elem.attributes[j]; | |
| if ( | |
| attr.name === "src" || | |
| attr.name === "srcset" || | |
| attr.name === "alt" | |
| ) { | |
| continue; | |
| } | |
| var copyTo = null; | |
| if (/\.(jpg|jpeg|png|webp)\s+\d/.test(attr.value)) { | |
| copyTo = "srcset"; | |
| } else if (/^\s*\S+\.(jpg|jpeg|png|webp)\S*\s*$/.test(attr.value)) { | |
| copyTo = "src"; | |
| } | |
| if (copyTo) { | |
| //if this is an img or picture, set the attribute directly | |
| if (elem.tagName === "IMG" || elem.tagName === "PICTURE") { | |
| elem.setAttribute(copyTo, attr.value); | |
| } else if ( | |
| elem.tagName === "FIGURE" && | |
| !this._getAllNodesWithTag(elem, ["img", "picture"]).length | |
| ) { | |
| //if the item is a <figure> that does not contain an image or picture, create one and place it inside the figure | |
| //see the nytimes-3 testcase for an example | |
| var img = this._doc.createElement("img"); | |
| img.setAttribute(copyTo, attr.value); | |
| elem.appendChild(img); | |
| } | |
| } | |
| } | |
| }, | |
| ); | |
| }, | |
| _getTextDensity: function (e, tags) { | |
| var textLength = this._getInnerText(e, true).length; | |
| if (textLength === 0) { | |
| return 0; | |
| } | |
| var childrenLength = 0; | |
| var children = this._getAllNodesWithTag(e, tags); | |
| this._forEachNode( | |
| children, | |
| (child) => (childrenLength += this._getInnerText(child, true).length), | |
| ); | |
| return childrenLength / textLength; | |
| }, | |
| /** | |
| * Clean an element of all tags of type "tag" if they look fishy. | |
| * "Fishy" is an algorithm based on content length, classnames, link density, number of images & embeds, etc. | |
| * | |
| * @return void | |
| **/ | |
| _cleanConditionally: function (e, tag) { | |
| if (!this._flagIsActive(this.FLAG_CLEAN_CONDITIONALLY)) return; | |
| // Gather counts for other typical elements embedded within. | |
| // Traverse backwards so we can remove nodes at the same time | |
| // without effecting the traversal. | |
| // | |
| // TODO: Consider taking into account original contentScore here. | |
| this._removeNodes(this._getAllNodesWithTag(e, [tag]), function (node) { | |
| // First check if this node IS data table, in which case don't remove it. | |
| var isDataTable = function (t) { | |
| return t._readabilityDataTable; | |
| }; | |
| var isList = tag === "ul" || tag === "ol"; | |
| if (!isList) { | |
| var listLength = 0; | |
| var listNodes = this._getAllNodesWithTag(node, ["ul", "ol"]); | |
| this._forEachNode( | |
| listNodes, | |
| (list) => (listLength += this._getInnerText(list).length), | |
| ); | |
| isList = listLength / this._getInnerText(node).length > 0.9; | |
| } | |
| if (tag === "table" && isDataTable(node)) { | |
| return false; | |
| } | |
| // Next check if we're inside a data table, in which case don't remove it as well. | |
| if (this._hasAncestorTag(node, "table", -1, isDataTable)) { | |
| return false; | |
| } | |
| if (this._hasAncestorTag(node, "code")) { | |
| return false; | |
| } | |
| var weight = this._getClassWeight(node); | |
| this.log("Cleaning Conditionally", node); | |
| var contentScore = 0; | |
| if (weight + contentScore < 0) { | |
| return true; | |
| } | |
| if (this._getCharCount(node, ",") < 10) { | |
| // If there are not very many commas, and the number of | |
| // non-paragraph elements is more than paragraphs or other | |
| // ominous signs, remove the element. | |
| var p = node.getElementsByTagName("p").length; | |
| var img = node.getElementsByTagName("img").length; | |
| var li = node.getElementsByTagName("li").length - 100; | |
| var input = node.getElementsByTagName("input").length; | |
| var headingDensity = this._getTextDensity(node, [ | |
| "h1", | |
| "h2", | |
| "h3", | |
| "h4", | |
| "h5", | |
| "h6", | |
| ]); | |
| var embedCount = 0; | |
| var embeds = this._getAllNodesWithTag(node, [ | |
| "object", | |
| "embed", | |
| "iframe", | |
| ]); | |
| for (var i = 0; i < embeds.length; i++) { | |
| // If this embed has attribute that matches video regex, don't delete it. | |
| for (var j = 0; j < embeds[i].attributes.length; j++) { | |
| if (this._allowedVideoRegex.test(embeds[i].attributes[j].value)) { | |
| return false; | |
| } | |
| } | |
| // For embed with <object> tag, check inner HTML as well. | |
| if ( | |
| embeds[i].tagName === "object" && | |
| this._allowedVideoRegex.test(embeds[i].innerHTML) | |
| ) { | |
| return false; | |
| } | |
| embedCount++; | |
| } | |
| var linkDensity = this._getLinkDensity(node); | |
| var contentLength = this._getInnerText(node).length; | |
| var haveToRemove = | |
| (img > 1 && p / img < 0.5 && !this._hasAncestorTag(node, "figure")) || | |
| (!isList && li > p) || | |
| input > Math.floor(p / 3) || | |
| (!isList && | |
| headingDensity < 0.9 && | |
| contentLength < 25 && | |
| (img === 0 || img > 2) && | |
| !this._hasAncestorTag(node, "figure")) || | |
| (!isList && weight < 25 && linkDensity > 0.2) || | |
| (weight >= 25 && linkDensity > 0.5) || | |
| (embedCount === 1 && contentLength < 75) || | |
| embedCount > 1; | |
| // Allow simple lists of images to remain in pages | |
| if (isList && haveToRemove) { | |
| for (var x = 0; x < node.children.length; x++) { | |
| let child = node.children[x]; | |
| // Don't filter in lists with li's that contain more than one child | |
| if (child.children.length > 1) { | |
| return haveToRemove; | |
| } | |
| } | |
| let li_count = node.getElementsByTagName("li").length; | |
| // Only allow the list to remain if every li contains an image | |
| if (img == li_count) { | |
| return false; | |
| } | |
| } | |
| return haveToRemove; | |
| } | |
| return false; | |
| }); | |
| }, | |
| /** | |
| * Clean out elements that match the specified conditions | |
| * | |
| * @param Element | |
| * @param Function determines whether a node should be removed | |
| * @return void | |
| **/ | |
| _cleanMatchedNodes: function (e, filter) { | |
| var endOfSearchMarkerNode = this._getNextNode(e, true); | |
| var next = this._getNextNode(e); | |
| while (next && next != endOfSearchMarkerNode) { | |
| if (filter.call(this, next, next.className + " " + next.id)) { | |
| next = this._removeAndGetNext(next); | |
| } else { | |
| next = this._getNextNode(next); | |
| } | |
| } | |
| }, | |
| /** | |
| * Clean out spurious headers from an Element. | |
| * | |
| * @param Element | |
| * @return void | |
| **/ | |
| _cleanHeaders: function (e) { | |
| let headingNodes = this._getAllNodesWithTag(e, ["h1", "h2"]); | |
| this._removeNodes(headingNodes, function (node) { | |
| let shouldRemove = this._getClassWeight(node) < 0; | |
| if (shouldRemove) { | |
| this.log("Removing header with low class weight:", node); | |
| } | |
| return shouldRemove; | |
| }); | |
| }, | |
| /** | |
| * Check if this node is an H1 or H2 element whose content is mostly | |
| * the same as the article title. | |
| * | |
| * @param Element the node to check. | |
| * @return boolean indicating whether this is a title-like header. | |
| */ | |
| _headerDuplicatesTitle: function (node) { | |
| if (node.tagName != "H1" && node.tagName != "H2") { | |
| return false; | |
| } | |
| var heading = this._getInnerText(node, false); | |
| this.log("Evaluating similarity of header:", heading, this._articleTitle); | |
| return this._textSimilarity(this._articleTitle, heading) > 0.75; | |
| }, | |
| _flagIsActive: function (flag) { | |
| return (this._flags & flag) > 0; | |
| }, | |
| _removeFlag: function (flag) { | |
| this._flags = this._flags & ~flag; | |
| }, | |
| _isProbablyVisible: function (node) { | |
| // Have to null-check node.style and node.className.indexOf to deal with SVG and MathML nodes. | |
| return ( | |
| (!node.style || node.style.display != "none") && | |
| !node.hasAttribute("hidden") && | |
| //check for "fallback-image" so that wikimedia math images are displayed | |
| (!node.hasAttribute("aria-hidden") || | |
| node.getAttribute("aria-hidden") != "true" || | |
| (node.className && | |
| node.className.indexOf && | |
| node.className.indexOf("fallback-image") !== -1)) | |
| ); | |
| }, | |
| /** | |
| * Runs readability. | |
| * | |
| * Workflow: | |
| * 1. Prep the document by removing script tags, css, etc. | |
| * 2. Build readability's DOM tree. | |
| * 3. Grab the article content from the current dom tree. | |
| * 4. Replace the current DOM tree with the new one. | |
| * 5. Read peacefully. | |
| * | |
| * @return void | |
| **/ | |
| parse: function () { | |
| // Avoid parsing too large documents, as per configuration option | |
| if (this._maxElemsToParse > 0) { | |
| var numTags = this._doc.getElementsByTagName("*").length; | |
| if (numTags > this._maxElemsToParse) { | |
| throw new Error( | |
| "Aborting parsing document; " + numTags + " elements found", | |
| ); | |
| } | |
| } | |
| // Unwrap image from noscript | |
| this._unwrapNoscriptImages(this._doc); | |
| // Extract JSON-LD metadata before removing scripts | |
| var jsonLd = this._disableJSONLD ? {} : this._getJSONLD(this._doc); | |
| // Remove script tags from the document. | |
| this._removeScripts(this._doc); | |
| this._prepDocument(); | |
| var metadata = this._getArticleMetadata(jsonLd); | |
| this._articleTitle = metadata.title; | |
| var articleContent = this._grabArticle(); | |
| if (!articleContent) return null; | |
| this.log("Grabbed: " + articleContent.innerHTML); | |
| this._postProcessContent(articleContent); | |
| // If we haven't found an excerpt in the article's metadata, use the article's | |
| // first paragraph as the excerpt. This is used for displaying a preview of | |
| // the article's content. | |
| if (!metadata.excerpt) { | |
| var paragraphs = articleContent.getElementsByTagName("p"); | |
| if (paragraphs.length > 0) { | |
| metadata.excerpt = paragraphs[0].textContent.trim(); | |
| } | |
| } | |
| var textContent = articleContent.textContent; | |
| return { | |
| title: this._articleTitle, | |
| byline: metadata.byline || this._articleByline, | |
| dir: this._articleDir, | |
| lang: this._articleLang, | |
| content: this._serializer(articleContent), | |
| textContent: textContent, | |
| length: textContent.length, | |
| excerpt: metadata.excerpt, | |
| siteName: metadata.siteName || this._articleSiteName, | |
| }; | |
| }, | |
| }; | |
| if (typeof module === "object") { | |
| /* global module */ | |
| module.exports = Readability; | |
| } | |