|
|
import { readdirSync } from "node:fs"; |
|
|
import { join } from "node:path"; |
|
|
import { docker } from "@dokploy/server/constants"; |
|
|
import { |
|
|
execAsync, |
|
|
execAsyncRemote, |
|
|
} from "@dokploy/server/utils/process/execAsync"; |
|
|
|
|
|
export interface IUpdateData { |
|
|
latestVersion: string | null; |
|
|
updateAvailable: boolean; |
|
|
} |
|
|
|
|
|
export const DEFAULT_UPDATE_DATA: IUpdateData = { |
|
|
latestVersion: null, |
|
|
updateAvailable: false, |
|
|
}; |
|
|
|
|
|
|
|
|
export const getDokployImageTag = () => { |
|
|
return process.env.RELEASE_TAG || "latest"; |
|
|
}; |
|
|
|
|
|
export const getDokployImage = () => { |
|
|
return `dokploy/dokploy:${getDokployImageTag()}`; |
|
|
}; |
|
|
|
|
|
export const pullLatestRelease = async () => { |
|
|
const stream = await docker.pull(getDokployImage()); |
|
|
await new Promise((resolve, reject) => { |
|
|
docker.modem.followProgress(stream, (err, res) => |
|
|
err ? reject(err) : resolve(res), |
|
|
); |
|
|
}); |
|
|
}; |
|
|
|
|
|
|
|
|
export const getServiceImageDigest = async () => { |
|
|
const { stdout } = await execAsync( |
|
|
"docker service inspect dokploy --format '{{.Spec.TaskTemplate.ContainerSpec.Image}}'", |
|
|
); |
|
|
|
|
|
const currentDigest = stdout.trim().split("@")[1]; |
|
|
|
|
|
if (!currentDigest) { |
|
|
throw new Error("Could not get current service image digest"); |
|
|
} |
|
|
|
|
|
return currentDigest; |
|
|
}; |
|
|
|
|
|
|
|
|
export const getUpdateData = async (): Promise<IUpdateData> => { |
|
|
let currentDigest: string; |
|
|
try { |
|
|
currentDigest = await getServiceImageDigest(); |
|
|
} catch { |
|
|
|
|
|
|
|
|
|
|
|
return DEFAULT_UPDATE_DATA; |
|
|
} |
|
|
|
|
|
const baseUrl = "https://hub.docker.com/v2/repositories/dokploy/dokploy/tags"; |
|
|
let url: string | null = `${baseUrl}?page_size=100`; |
|
|
let allResults: { digest: string; name: string }[] = []; |
|
|
while (url) { |
|
|
const response = await fetch(url, { |
|
|
method: "GET", |
|
|
headers: { "Content-Type": "application/json" }, |
|
|
}); |
|
|
|
|
|
const data = (await response.json()) as { |
|
|
next: string | null; |
|
|
results: { digest: string; name: string }[]; |
|
|
}; |
|
|
|
|
|
allResults = allResults.concat(data.results); |
|
|
url = data?.next; |
|
|
} |
|
|
|
|
|
const imageTag = getDokployImageTag(); |
|
|
const searchedDigest = allResults.find((t) => t.name === imageTag)?.digest; |
|
|
|
|
|
if (!searchedDigest) { |
|
|
return DEFAULT_UPDATE_DATA; |
|
|
} |
|
|
|
|
|
if (imageTag === "latest") { |
|
|
const versionedTag = allResults.find( |
|
|
(t) => t.digest === searchedDigest && t.name.startsWith("v"), |
|
|
); |
|
|
|
|
|
if (!versionedTag) { |
|
|
return DEFAULT_UPDATE_DATA; |
|
|
} |
|
|
|
|
|
const { name: latestVersion, digest } = versionedTag; |
|
|
const updateAvailable = digest !== currentDigest; |
|
|
|
|
|
return { latestVersion, updateAvailable }; |
|
|
} |
|
|
const updateAvailable = searchedDigest !== currentDigest; |
|
|
return { latestVersion: imageTag, updateAvailable }; |
|
|
}; |
|
|
|
|
|
interface TreeDataItem { |
|
|
id: string; |
|
|
name: string; |
|
|
type: "file" | "directory"; |
|
|
children?: TreeDataItem[]; |
|
|
} |
|
|
|
|
|
export const readDirectory = async ( |
|
|
dirPath: string, |
|
|
serverId?: string, |
|
|
): Promise<TreeDataItem[]> => { |
|
|
if (serverId) { |
|
|
const { stdout } = await execAsyncRemote( |
|
|
serverId, |
|
|
` |
|
|
process_items() { |
|
|
local parent_dir="$1" |
|
|
local __resultvar=$2 |
|
|
|
|
|
local items_json="" |
|
|
local first=true |
|
|
for item in "$parent_dir"/*; do |
|
|
[ -e "$item" ] || continue |
|
|
process_item "$item" item_json |
|
|
if [ "$first" = true ]; then |
|
|
first=false |
|
|
items_json="$item_json" |
|
|
else |
|
|
items_json="$items_json,$item_json" |
|
|
fi |
|
|
done |
|
|
|
|
|
eval $__resultvar="'[$items_json]'" |
|
|
} |
|
|
|
|
|
process_item() { |
|
|
local item_path="$1" |
|
|
local __resultvar=$2 |
|
|
|
|
|
local item_name=$(basename "$item_path") |
|
|
local escaped_name=$(echo "$item_name" | sed 's/"/\\"/g') |
|
|
local escaped_path=$(echo "$item_path" | sed 's/"/\\"/g') |
|
|
|
|
|
if [ -d "$item_path" ]; then |
|
|
# Is directory |
|
|
process_items "$item_path" children_json |
|
|
local json='{"id":"'"$escaped_path"'","name":"'"$escaped_name"'","type":"directory","children":'"$children_json"'}' |
|
|
else |
|
|
# Is file |
|
|
local json='{"id":"'"$escaped_path"'","name":"'"$escaped_name"'","type":"file"}' |
|
|
fi |
|
|
|
|
|
eval $__resultvar="'$json'" |
|
|
} |
|
|
|
|
|
root_dir=${dirPath} |
|
|
|
|
|
process_items "$root_dir" json_output |
|
|
|
|
|
echo "$json_output" |
|
|
`, |
|
|
); |
|
|
const result = JSON.parse(stdout); |
|
|
return result; |
|
|
} |
|
|
|
|
|
const stack = [dirPath]; |
|
|
const result: TreeDataItem[] = []; |
|
|
const parentMap: Record<string, TreeDataItem[]> = {}; |
|
|
|
|
|
while (stack.length > 0) { |
|
|
const currentPath = stack.pop(); |
|
|
if (!currentPath) continue; |
|
|
|
|
|
const items = readdirSync(currentPath, { withFileTypes: true }); |
|
|
const currentDirectoryResult: TreeDataItem[] = []; |
|
|
|
|
|
for (const item of items) { |
|
|
const fullPath = join(currentPath, item.name); |
|
|
if (item.isDirectory()) { |
|
|
stack.push(fullPath); |
|
|
const directoryItem: TreeDataItem = { |
|
|
id: fullPath, |
|
|
name: item.name, |
|
|
type: "directory", |
|
|
children: [], |
|
|
}; |
|
|
currentDirectoryResult.push(directoryItem); |
|
|
parentMap[fullPath] = directoryItem.children as TreeDataItem[]; |
|
|
} else { |
|
|
const fileItem: TreeDataItem = { |
|
|
id: fullPath, |
|
|
name: item.name, |
|
|
type: "file", |
|
|
}; |
|
|
currentDirectoryResult.push(fileItem); |
|
|
} |
|
|
} |
|
|
|
|
|
if (parentMap[currentPath]) { |
|
|
parentMap[currentPath].push(...currentDirectoryResult); |
|
|
} else { |
|
|
result.push(...currentDirectoryResult); |
|
|
} |
|
|
} |
|
|
return result; |
|
|
}; |
|
|
|
|
|
export const cleanupFullDocker = async (serverId?: string | null) => { |
|
|
const cleanupImages = "docker image prune --force"; |
|
|
const cleanupVolumes = "docker volume prune --force"; |
|
|
const cleanupContainers = "docker container prune --force"; |
|
|
const cleanupSystem = "docker system prune --force --volumes"; |
|
|
const cleanupBuilder = "docker builder prune --force"; |
|
|
|
|
|
try { |
|
|
if (serverId) { |
|
|
await execAsyncRemote( |
|
|
serverId, |
|
|
` |
|
|
${cleanupImages} |
|
|
${cleanupVolumes} |
|
|
${cleanupContainers} |
|
|
${cleanupSystem} |
|
|
${cleanupBuilder} |
|
|
`, |
|
|
); |
|
|
} |
|
|
await execAsync(` |
|
|
${cleanupImages} |
|
|
${cleanupVolumes} |
|
|
${cleanupContainers} |
|
|
${cleanupSystem} |
|
|
${cleanupBuilder} |
|
|
`); |
|
|
} catch (error) { |
|
|
console.log(error); |
|
|
} |
|
|
}; |
|
|
|