| | import fs from 'fs/promises' |
| | import path from 'path' |
| | import unified from 'unified' |
| | import markdown from 'remark-parse' |
| | import remarkToRehype from 'remark-rehype' |
| | import raw from 'rehype-raw' |
| | import visit from 'unist-util-visit' |
| | import GithubSlugger from 'github-slugger' |
| | import matter from 'gray-matter' |
| | import * as github from '@actions/github' |
| | import { setFailed } from '@actions/core' |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | interface Document { |
| | body: string |
| | path: string |
| | headings: string[] |
| | source?: string |
| | related?: { |
| | links: string[] |
| | } |
| | } |
| |
|
| | interface Errors { |
| | doc: Document |
| | link: string[] |
| | hash: string[] |
| | source: string[] |
| | related: string[] |
| | } |
| |
|
| | type ErrorType = Exclude<keyof Errors, 'doc'> |
| |
|
| | interface Comment { |
| | id: number |
| | } |
| |
|
| | const DOCS_PATH = '/docs/' |
| | const ERRORS_PATH = '/errors/' |
| | const EXCLUDED_HASHES = ['top'] |
| | const COMMENT_TAG = '<!-- LINK_CHECKER_COMMENT -->' |
| |
|
| | const { context, getOctokit } = github |
| | const octokit = getOctokit(process.env.GITHUB_TOKEN!) |
| | const { owner, repo } = context.repo |
| | const pullRequest = context.payload.pull_request |
| | if (!pullRequest) { |
| | console.log('Skipping since this is not a pull request') |
| | process.exit(0) |
| | } |
| | const sha = pullRequest.head.sha |
| | const isFork = pullRequest.head.repo.fork |
| | const prNumber = pullRequest.number |
| |
|
| | const slugger = new GithubSlugger() |
| |
|
| | |
| | async function getAllMdxFilePaths( |
| | directoriesToScan: string[], |
| | fileList: string[] = [] |
| | ): Promise<string[]> { |
| | for (const dir of directoriesToScan) { |
| | const dirPath = path.join('.', dir) |
| | const files = await fs.readdir(dirPath) |
| | for (const file of files) { |
| | const filePath = path.join(dirPath, file) |
| | const stats = await fs.stat(filePath) |
| | if (stats.isDirectory()) { |
| | fileList = await getAllMdxFilePaths([filePath], fileList) |
| | } else if (path.extname(file) === '.mdx') { |
| | fileList.push(filePath) |
| | } |
| | } |
| | } |
| |
|
| | return fileList |
| | } |
| |
|
| | |
| | function getHeadingsFromMarkdownTree( |
| | tree: ReturnType<typeof markdownProcessor.parse> |
| | ): string[] { |
| | const headings: string[] = [] |
| | slugger.reset() |
| |
|
| | visit(tree, 'heading', (node) => { |
| | let headingText = '' |
| | |
| | |
| | visit(node, (node: any) => { |
| | if (node.value) { |
| | headingText += node.value |
| | } |
| | }) |
| | headings.push(slugger.slug(headingText)) |
| | }) |
| |
|
| | return headings |
| | } |
| |
|
| | |
| | const markdownProcessor = unified() |
| | .use(markdown) |
| | .use(remarkToRehype, { allowDangerousHTML: true }) |
| | .use(raw) |
| | .use(function compiler() { |
| | |
| | |
| | |
| | this.Compiler = function treeCompiler(tree) { |
| | return tree |
| | } |
| | }) |
| |
|
| | |
| | function normalizePath(filePath: string): string { |
| | if (filePath.startsWith(ERRORS_PATH.substring(1))) { |
| | return ( |
| | filePath |
| | |
| | |
| | .replace(ERRORS_PATH.substring(1), DOCS_PATH.substring(1) + 'messages/') |
| | .replace('.mdx', '') |
| | ) |
| | } |
| |
|
| | return ( |
| | |
| | |
| | filePath |
| | |
| | |
| | |
| | |
| | |
| | |
| | .replace(DOCS_PATH.substring(1), '') |
| | |
| | .replace(/(\d\d-)/g, '') |
| | .replace('.mdx', '') |
| | .replace('/index', '') |
| | ) |
| | } |
| |
|
| | |
| | let documentMap: Map<string, Document> |
| |
|
| | |
| | |
| | |
| | |
| | |
| | async function prepareDocumentMapEntry( |
| | filePath: string |
| | ): Promise<[string, Document]> { |
| | try { |
| | const mdxContent = await fs.readFile(filePath, 'utf8') |
| | const { content, data } = matter(mdxContent) |
| | const tree = markdownProcessor.parse(content) |
| | const headings = getHeadingsFromMarkdownTree(tree) |
| | const normalizedUrlPath = normalizePath(filePath) |
| |
|
| | return [ |
| | normalizedUrlPath, |
| | { body: content, path: filePath, headings, ...data }, |
| | ] |
| | } catch (error) { |
| | setFailed(`Error preparing document map for file ${filePath}: ${error}`) |
| | return ['', {} as Document] |
| | } |
| | } |
| |
|
| | |
| | function validateInternalLink(errors: Errors, href: string): void { |
| | |
| | const [link, hash] = href.replace(DOCS_PATH, '').split('#', 2) |
| |
|
| | let foundPage |
| |
|
| | if (link.startsWith('messages/')) { |
| | |
| | |
| | foundPage = documentMap.get(DOCS_PATH.substring(1) + link) |
| | } else { |
| | |
| | |
| | foundPage = documentMap.get(link) |
| | } |
| |
|
| | if (!foundPage) { |
| | errors.link.push(href) |
| | } else if (hash && !EXCLUDED_HASHES.includes(hash)) { |
| | |
| | const foundPageSource = foundPage.source |
| | ? documentMap.get(foundPage.source) |
| | : undefined |
| |
|
| | |
| | const hashFound = (foundPageSource || foundPage).headings.includes(hash) |
| |
|
| | if (!hashFound) { |
| | errors.hash.push(href) |
| | } |
| | } |
| | } |
| |
|
| | |
| | function validateHashLink(errors: Errors, href: string, doc: Document): void { |
| | const hashLink = href.replace('#', '') |
| |
|
| | if (!EXCLUDED_HASHES.includes(hashLink) && !doc.headings.includes(hashLink)) { |
| | errors.hash.push(href) |
| | } |
| | } |
| |
|
| | |
| | function validateSourceLinks(doc: Document, errors: Errors): void { |
| | if (doc.source && !documentMap.get(doc.source)) { |
| | errors.source.push(doc.source) |
| | } |
| | } |
| |
|
| | |
| | function validateRelatedLinks(doc: Document, errors: Errors): void { |
| | if (doc.related && doc.related.links) { |
| | doc.related.links.forEach((link) => { |
| | if (!documentMap.get(link)) { |
| | errors.related.push(link) |
| | } |
| | }) |
| | } |
| | } |
| |
|
| | |
| | function traverseTreeAndValidateLinks(tree: any, doc: Document): Errors { |
| | const errors: Errors = { |
| | doc, |
| | link: [], |
| | hash: [], |
| | source: [], |
| | related: [], |
| | } |
| |
|
| | try { |
| | visit(tree, (node: any) => { |
| | if (node.type === 'element' && node.tagName === 'a') { |
| | const href = node.properties.href |
| |
|
| | if (!href) return |
| |
|
| | if (href.startsWith(DOCS_PATH)) { |
| | validateInternalLink(errors, href) |
| | } else if (href.startsWith('#')) { |
| | validateHashLink(errors, href, doc) |
| | } |
| | } |
| | }) |
| |
|
| | validateSourceLinks(doc, errors) |
| | validateRelatedLinks(doc, errors) |
| | } catch (error) { |
| | setFailed('Error traversing tree: ' + error) |
| | } |
| |
|
| | return errors |
| | } |
| |
|
| | async function findBotComment(): Promise<Comment | undefined> { |
| | try { |
| | const { data: comments } = await octokit.rest.issues.listComments({ |
| | owner, |
| | repo, |
| | issue_number: prNumber, |
| | }) |
| |
|
| | return comments.find((c) => c.body?.includes(COMMENT_TAG)) |
| | } catch (error) { |
| | setFailed('Error finding bot comment: ' + error) |
| | return undefined |
| | } |
| | } |
| |
|
| | async function updateComment( |
| | comment: string, |
| | botComment: Comment |
| | ): Promise<string> { |
| | try { |
| | const { data } = await octokit.rest.issues.updateComment({ |
| | owner, |
| | repo, |
| | comment_id: botComment.id, |
| | body: comment, |
| | }) |
| |
|
| | return data.html_url |
| | } catch (error) { |
| | setFailed('Error updating comment: ' + error) |
| | return '' |
| | } |
| | } |
| |
|
| | async function createComment(comment: string): Promise<string> { |
| | if (isFork) { |
| | setFailed( |
| | 'The action could not create a Github comment because it is initiated from a forked repo. View the action logs for a list of broken links.' |
| | ) |
| |
|
| | return '' |
| | } else { |
| | try { |
| | const { data } = await octokit.rest.issues.createComment({ |
| | owner, |
| | repo, |
| | issue_number: prNumber, |
| | body: comment, |
| | }) |
| |
|
| | return data.html_url |
| | } catch (error) { |
| | setFailed('Error creating comment: ' + error) |
| | return '' |
| | } |
| | } |
| | } |
| |
|
| | const formatTableRow = ( |
| | link: string, |
| | errorType: ErrorType, |
| | docPath: string |
| | ) => { |
| | return `| ${link} | ${errorType} | [/${docPath}](https://github.com/vercel/next.js/blob/${sha}/${docPath}) | \n` |
| | } |
| |
|
| | async function updateCheckStatus( |
| | errorsExist: boolean, |
| | commentUrl?: string |
| | ): Promise<void> { |
| | const checkName = 'Docs Link Validation' |
| |
|
| | let summary, text |
| |
|
| | if (errorsExist) { |
| | summary = |
| | 'This PR introduces broken links to the docs. Click details for a list.' |
| | text = `[See the comment for details](${commentUrl})` |
| | } else { |
| | summary = 'No broken links found' |
| | } |
| |
|
| | const checkParams = { |
| | owner, |
| | repo, |
| | name: checkName, |
| | head_sha: sha, |
| | status: 'completed', |
| | conclusion: errorsExist ? 'failure' : 'success', |
| | output: { |
| | title: checkName, |
| | summary: summary, |
| | text: text, |
| | }, |
| | } |
| |
|
| | if (isFork) { |
| | if (errorsExist) { |
| | setFailed( |
| | 'This PR introduces broken links to the docs. The action could not create a Github check because it is initiated from a forked repo.' |
| | ) |
| | } else { |
| | console.log('Link validation was successful.') |
| | } |
| | } else { |
| | try { |
| | await octokit.rest.checks.create(checkParams) |
| | } catch (error) { |
| | setFailed('Failed to create check: ' + error) |
| | } |
| | } |
| | } |
| |
|
| | |
| | async function validateAllInternalLinks(): Promise<void> { |
| | try { |
| | const allMdxFilePaths = await getAllMdxFilePaths([DOCS_PATH, ERRORS_PATH]) |
| |
|
| | documentMap = new Map( |
| | await Promise.all(allMdxFilePaths.map(prepareDocumentMapEntry)) |
| | ) |
| |
|
| | const docProcessingPromises = allMdxFilePaths.map(async (filePath) => { |
| | const doc = documentMap.get(normalizePath(filePath)) |
| | if (doc) { |
| | const tree = (await markdownProcessor.process(doc.body)).contents |
| | return traverseTreeAndValidateLinks(tree, doc) |
| | } else { |
| | return { |
| | doc: {} as Document, |
| | link: [], |
| | hash: [], |
| | source: [], |
| | related: [], |
| | } as Errors |
| | } |
| | }) |
| |
|
| | const allErrors = await Promise.all(docProcessingPromises) |
| |
|
| | let errorsExist = false |
| |
|
| | let errorRows: string[] = [] |
| |
|
| | const errorTypes: ErrorType[] = ['link', 'hash', 'source', 'related'] |
| | allErrors.forEach((errors) => { |
| | const { |
| | doc: { path: docPath }, |
| | } = errors |
| |
|
| | errorTypes.forEach((errorType) => { |
| | if (errors[errorType].length > 0) { |
| | errorsExist = true |
| | errors[errorType].forEach((link) => { |
| | errorRows.push(formatTableRow(link, errorType, docPath)) |
| | }) |
| | } |
| | }) |
| | }) |
| |
|
| | const errorComment = [ |
| | 'Hi there :wave:\n\nIt looks like this PR introduces broken links to the docs, please take a moment to fix them before merging:\n\n| Broken link | Type | File | \n| ----------- | ----------- | ----------- | \n', |
| | ...errorRows, |
| | '\nThank you :pray:', |
| | ].join('') |
| |
|
| | const botComment = await findBotComment() |
| |
|
| | let commentUrl |
| |
|
| | if (errorsExist) { |
| | const comment = `${COMMENT_TAG}\n${errorComment}` |
| | if (botComment) { |
| | commentUrl = await updateComment(comment, botComment) |
| | } else { |
| | commentUrl = await createComment(comment) |
| | } |
| |
|
| | const errorTableData = allErrors.flatMap((errors) => { |
| | const { doc } = errors |
| |
|
| | return errorTypes.flatMap((errorType) => |
| | errors[errorType].map((link) => ({ |
| | docPath: doc.path, |
| | errorType, |
| | link, |
| | })) |
| | ) |
| | }) |
| |
|
| | console.log('This PR introduces broken links to the docs:') |
| | console.table(errorTableData, ['link', 'type', 'docPath']) |
| | } else if (botComment) { |
| | const comment = `${COMMENT_TAG}\nAll broken links are now fixed, thank you!` |
| | commentUrl = await updateComment(comment, botComment) |
| | } |
| |
|
| | try { |
| | await updateCheckStatus(errorsExist, commentUrl) |
| | } catch (error) { |
| | setFailed('Failed to create Github check: ' + error) |
| | } |
| | } catch (error) { |
| | setFailed('Error validating internal links: ' + error) |
| | } |
| | } |
| |
|
| | validateAllInternalLinks() |
| |
|