| {"repo_name": "yap", "file_name": "/yap/Sources/yap/OutputFormat.swift", "inference_info": {"prefix_code": "import ArgumentParser\nimport CoreMedia\nimport Foundation\n\nenum OutputFormat: String, EnumerableFlag {\n case txt\n case srt\n\n // MARK: Internal\n\n var needsAudioTimeRange: Bool {\n switch self {\n case .srt: true\n default: false\n }\n }\n\n ", "suffix_code": "\n}\n", "middle_code": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }", "code_description": null, "fill_type": "FUNCTION_TYPE", "language_type": "swift", "sub_task_type": null}, "context_code": [["/yap/Sources/yap/Extensions/AttributedString+Extensions.swift", "import CoreMedia\nimport Foundation\nimport NaturalLanguage\n\nextension AttributedString {\n func sentences(maxLength: Int? = nil) -> [AttributedString] {\n let tokenizer = NLTokenizer(unit: .sentence)\n let string = String(characters)\n tokenizer.string = string\n let sentenceRanges = tokenizer.tokens(for: string.startIndex..<string.endIndex).map {\n (\n $0,\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n )\n }\n let ranges = sentenceRanges.flatMap { sentenceStringRange, sentenceRange in\n let sentence = self[sentenceRange]\n guard let maxLength, sentence.characters.count > maxLength else {\n return [sentenceRange]\n }\n\n let wordTokenizer = NLTokenizer(unit: .word)\n wordTokenizer.string = string\n var wordRanges = wordTokenizer.tokens(for: sentenceStringRange).map {\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n }\n guard !wordRanges.isEmpty else { return [sentenceRange] }\n wordRanges[0] = sentenceRange.lowerBound..<wordRanges[0].upperBound\n wordRanges[wordRanges.count - 1] = wordRanges[wordRanges.count - 1].lowerBound..<sentenceRange.upperBound\n\n var ranges: [Range<AttributedString.Index>] = []\n for wordRange in wordRanges {\n if let lastRange = ranges.last,\n self[lastRange].characters.count + self[wordRange].characters.count <= maxLength {\n ranges[ranges.count - 1] = lastRange.lowerBound..<wordRange.upperBound\n } else {\n ranges.append(wordRange)\n }\n }\n\n return ranges\n }\n\n return ranges.compactMap { range in\n let audioTimeRanges = self[range].runs.filter {\n !String(self[$0.range].characters)\n .trimmingCharacters(in: .whitespacesAndNewlines).isEmpty\n }.compactMap(\\.audioTimeRange)\n guard !audioTimeRanges.isEmpty else { return nil }\n let start = audioTimeRanges.first!.start\n let end = audioTimeRanges.last!.end\n var attributes = AttributeContainer()\n attributes[AttributeScopes.SpeechAttributes.TimeRangeAttribute.self] = CMTimeRange(\n start: start,\n end: end\n )\n return AttributedString(self[range].characters, attributes: attributes)\n }\n }\n}\n"], ["/yap/Sources/yap/Transcribe.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - Transcribe\n\n@MainActor struct Transcribe: AsyncParsableCommand {\n @Option(\n name: .shortAndLong,\n help: \"(default: current)\",\n transform: Locale.init(identifier:)\n ) var locale: Locale = .init(identifier: Locale.current.identifier)\n\n @Flag(\n help: \"Replaces certain words and phrases with a redacted form.\"\n ) var censor: Bool = false\n\n @Argument(\n help: \"Path to an audio or video file to transcribe.\",\n transform: URL.init(fileURLWithPath:)\n ) var inputFile: URL\n\n @Flag(\n help: \"Output format for the transcription.\",\n ) var outputFormat: OutputFormat = .txt\n\n @Option(\n name: .shortAndLong,\n help: \"Path to save the transcription output. If not provided, output will be printed to stdout.\",\n transform: URL.init(fileURLWithPath:)\n ) var outputFile: URL?\n\n mutating func run() async throws {\n let piped = isatty(STDOUT_FILENO) == 0\n struct DevNull: StandardPipelining { func write(content _: String) {} }\n let noora = if piped {\n Noora(standardPipelines: .init(output: DevNull()))\n } else {\n Noora()\n }\n\n let supported = await SpeechTranscriber.supportedLocales\n guard supported.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) else {\n noora.error(.alert(\"Locale \\\"\\(locale.identifier)\\\" is not supported. Supported locales:\\n\\(supported.map(\\.identifier))\"))\n throw Error.unsupportedLocale\n }\n\n for locale in await AssetInventory.allocatedLocales {\n await AssetInventory.deallocate(locale: locale)\n }\n try await AssetInventory.allocate(locale: locale)\n\n let transcriber = SpeechTranscriber(\n locale: locale,\n transcriptionOptions: censor ? [.etiquetteReplacements] : [],\n reportingOptions: [],\n attributeOptions: outputFormat.needsAudioTimeRange ? [.audioTimeRange] : []\n )\n let modules: [any SpeechModule] = [transcriber]\n let installed = await Set(SpeechTranscriber.installedLocales)\n if !installed.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) {\n if let request = try await AssetInventory.assetInstallationRequest(supporting: modules) {\n try await noora.progressBarStep(\n message: \"Downloading required assets…\"\n ) { @Sendable progressCallback in\n struct ProgressCallback: @unchecked Sendable {\n let callback: (Double) -> Void\n }\n let progressCallback = ProgressCallback(callback: progressCallback)\n Task {\n while !request.progress.isFinished {\n progressCallback.callback(request.progress.fractionCompleted)\n try? await Task.sleep(for: .seconds(0.1))\n }\n }\n try await request.downloadAndInstall()\n }\n }\n }\n\n let analyzer = SpeechAnalyzer(modules: modules)\n\n let audioFile = try AVAudioFile(forReading: inputFile)\n let audioFileDuration: TimeInterval = Double(audioFile.length) / audioFile.processingFormat.sampleRate\n try await analyzer.start(inputAudioFile: audioFile, finishAfterFile: true)\n\n var transcript: AttributedString = \"\"\n\n var w = winsize()\n let terminalColumns = if ioctl(STDOUT_FILENO, UInt(TIOCGWINSZ), &w) == 0 {\n max(Int(w.ws_col), 9)\n } else { 64 }\n\n try await noora.progressStep(\n message: \"Transcribing audio using locale: \\\"\\(locale.identifier)\\\"…\",\n successMessage: \"Audio transcribed using locale: \\\"\\(locale.identifier)\\\"\",\n errorMessage: \"Failed to transcribe audio using locale: \\\"\\(locale.identifier)\\\"\",\n showSpinner: true\n ) { @Sendable progressHandler in\n for try await result in transcriber.results {\n await MainActor.run {\n transcript += result.text\n }\n let progress = max(min(result.resultsFinalizationTime.seconds / audioFileDuration, 1), 0)\n var percent = progress.formatted(.percent.precision(.fractionLength(0)))\n let oneHundredPercent = 1.0.formatted(.percent.precision(.fractionLength(0)))\n percent = String(String(repeating: \" \", count: max(oneHundredPercent.count - percent.count, 0))) + percent\n let message = \"[\\(percent)] \\(String(result.text.characters).trimmingCharacters(in: .whitespaces).prefix(terminalColumns - \"⠋ [\\(oneHundredPercent)] \".count))\"\n progressHandler(message)\n }\n }\n\n if let outputFile {\n try outputFormat.text(for: transcript).write(\n to: outputFile,\n atomically: false,\n encoding: .utf8\n )\n noora.success(.alert(\"Transcription written to \\(outputFile.path)\"))\n }\n\n if piped || outputFile == nil {\n print(outputFormat.text(for: transcript))\n }\n }\n}\n\n// MARK: Transcribe.Error\n\nextension Transcribe {\n enum Error: Swift.Error {\n case unsupportedLocale\n }\n}\n"], ["/yap/Sources/yap/Yap.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - yap\n\n@main struct Yap: AsyncParsableCommand {\n static let configuration = CommandConfiguration(\n abstract: \"A CLI for on-device speech transcription.\",\n subcommands: [\n Transcribe.self\n ],\n defaultSubcommand: Transcribe.self\n )\n}\n"]], "task_instance_info": {"created_time": "2025-08-21 03:27:54", "created_task_model": "DeepSeek-R1", "function_skeleton": "func text ( for transcript: AttributedString ) -> String { }", "code_description": "This code segment functions to: Convert structured transcript data into formatted text output based on specified format requirements, handling both plain text conversion and timed subtitle generation with precise synchronization metadata."}, "inference_content": {"inference_model": "DeepSeek-R1", "inference_result": "```swift\n[TASK_BEGIN]\n func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n return String(transcript.characters)\n }\n }\n[TASK_END]\n```", "inference_time": "2025-08-21 03-28-32"}, "editdistance_info": {"edit_distance": 14.1914, "calculate_time": "2025-08-21 03:28:32", "true_code_clean": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }", "predict_code_clean": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n return String(transcript.characters)\n }\n }"}} | |
| {"repo_name": "yap", "file_name": "/yap/Sources/yap/OutputFormat.swift", "inference_info": {"prefix_code": "import ArgumentParser\nimport CoreMedia\nimport Foundation\n\nenum OutputFormat: String, EnumerableFlag {\n case txt\n case srt\n\n // MARK: Internal\n\n var needsAudioTimeRange: Bool {\n switch self {\n case .srt: true\n default: false\n }\n }\n\n ", "suffix_code": "\n}\n", "middle_code": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }", "code_description": null, "fill_type": "FUNCTION_TYPE", "language_type": "swift", "sub_task_type": null}, "context_code": [["/yap/Sources/yap/Extensions/AttributedString+Extensions.swift", "import CoreMedia\nimport Foundation\nimport NaturalLanguage\n\nextension AttributedString {\n func sentences(maxLength: Int? = nil) -> [AttributedString] {\n let tokenizer = NLTokenizer(unit: .sentence)\n let string = String(characters)\n tokenizer.string = string\n let sentenceRanges = tokenizer.tokens(for: string.startIndex..<string.endIndex).map {\n (\n $0,\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n )\n }\n let ranges = sentenceRanges.flatMap { sentenceStringRange, sentenceRange in\n let sentence = self[sentenceRange]\n guard let maxLength, sentence.characters.count > maxLength else {\n return [sentenceRange]\n }\n\n let wordTokenizer = NLTokenizer(unit: .word)\n wordTokenizer.string = string\n var wordRanges = wordTokenizer.tokens(for: sentenceStringRange).map {\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n }\n guard !wordRanges.isEmpty else { return [sentenceRange] }\n wordRanges[0] = sentenceRange.lowerBound..<wordRanges[0].upperBound\n wordRanges[wordRanges.count - 1] = wordRanges[wordRanges.count - 1].lowerBound..<sentenceRange.upperBound\n\n var ranges: [Range<AttributedString.Index>] = []\n for wordRange in wordRanges {\n if let lastRange = ranges.last,\n self[lastRange].characters.count + self[wordRange].characters.count <= maxLength {\n ranges[ranges.count - 1] = lastRange.lowerBound..<wordRange.upperBound\n } else {\n ranges.append(wordRange)\n }\n }\n\n return ranges\n }\n\n return ranges.compactMap { range in\n let audioTimeRanges = self[range].runs.filter {\n !String(self[$0.range].characters)\n .trimmingCharacters(in: .whitespacesAndNewlines).isEmpty\n }.compactMap(\\.audioTimeRange)\n guard !audioTimeRanges.isEmpty else { return nil }\n let start = audioTimeRanges.first!.start\n let end = audioTimeRanges.last!.end\n var attributes = AttributeContainer()\n attributes[AttributeScopes.SpeechAttributes.TimeRangeAttribute.self] = CMTimeRange(\n start: start,\n end: end\n )\n return AttributedString(self[range].characters, attributes: attributes)\n }\n }\n}\n"], ["/yap/Sources/yap/Transcribe.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - Transcribe\n\n@MainActor struct Transcribe: AsyncParsableCommand {\n @Option(\n name: .shortAndLong,\n help: \"(default: current)\",\n transform: Locale.init(identifier:)\n ) var locale: Locale = .init(identifier: Locale.current.identifier)\n\n @Flag(\n help: \"Replaces certain words and phrases with a redacted form.\"\n ) var censor: Bool = false\n\n @Argument(\n help: \"Path to an audio or video file to transcribe.\",\n transform: URL.init(fileURLWithPath:)\n ) var inputFile: URL\n\n @Flag(\n help: \"Output format for the transcription.\",\n ) var outputFormat: OutputFormat = .txt\n\n @Option(\n name: .shortAndLong,\n help: \"Path to save the transcription output. If not provided, output will be printed to stdout.\",\n transform: URL.init(fileURLWithPath:)\n ) var outputFile: URL?\n\n mutating func run() async throws {\n let piped = isatty(STDOUT_FILENO) == 0\n struct DevNull: StandardPipelining { func write(content _: String) {} }\n let noora = if piped {\n Noora(standardPipelines: .init(output: DevNull()))\n } else {\n Noora()\n }\n\n let supported = await SpeechTranscriber.supportedLocales\n guard supported.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) else {\n noora.error(.alert(\"Locale \\\"\\(locale.identifier)\\\" is not supported. Supported locales:\\n\\(supported.map(\\.identifier))\"))\n throw Error.unsupportedLocale\n }\n\n for locale in await AssetInventory.allocatedLocales {\n await AssetInventory.deallocate(locale: locale)\n }\n try await AssetInventory.allocate(locale: locale)\n\n let transcriber = SpeechTranscriber(\n locale: locale,\n transcriptionOptions: censor ? [.etiquetteReplacements] : [],\n reportingOptions: [],\n attributeOptions: outputFormat.needsAudioTimeRange ? [.audioTimeRange] : []\n )\n let modules: [any SpeechModule] = [transcriber]\n let installed = await Set(SpeechTranscriber.installedLocales)\n if !installed.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) {\n if let request = try await AssetInventory.assetInstallationRequest(supporting: modules) {\n try await noora.progressBarStep(\n message: \"Downloading required assets…\"\n ) { @Sendable progressCallback in\n struct ProgressCallback: @unchecked Sendable {\n let callback: (Double) -> Void\n }\n let progressCallback = ProgressCallback(callback: progressCallback)\n Task {\n while !request.progress.isFinished {\n progressCallback.callback(request.progress.fractionCompleted)\n try? await Task.sleep(for: .seconds(0.1))\n }\n }\n try await request.downloadAndInstall()\n }\n }\n }\n\n let analyzer = SpeechAnalyzer(modules: modules)\n\n let audioFile = try AVAudioFile(forReading: inputFile)\n let audioFileDuration: TimeInterval = Double(audioFile.length) / audioFile.processingFormat.sampleRate\n try await analyzer.start(inputAudioFile: audioFile, finishAfterFile: true)\n\n var transcript: AttributedString = \"\"\n\n var w = winsize()\n let terminalColumns = if ioctl(STDOUT_FILENO, UInt(TIOCGWINSZ), &w) == 0 {\n max(Int(w.ws_col), 9)\n } else { 64 }\n\n try await noora.progressStep(\n message: \"Transcribing audio using locale: \\\"\\(locale.identifier)\\\"…\",\n successMessage: \"Audio transcribed using locale: \\\"\\(locale.identifier)\\\"\",\n errorMessage: \"Failed to transcribe audio using locale: \\\"\\(locale.identifier)\\\"\",\n showSpinner: true\n ) { @Sendable progressHandler in\n for try await result in transcriber.results {\n await MainActor.run {\n transcript += result.text\n }\n let progress = max(min(result.resultsFinalizationTime.seconds / audioFileDuration, 1), 0)\n var percent = progress.formatted(.percent.precision(.fractionLength(0)))\n let oneHundredPercent = 1.0.formatted(.percent.precision(.fractionLength(0)))\n percent = String(String(repeating: \" \", count: max(oneHundredPercent.count - percent.count, 0))) + percent\n let message = \"[\\(percent)] \\(String(result.text.characters).trimmingCharacters(in: .whitespaces).prefix(terminalColumns - \"⠋ [\\(oneHundredPercent)] \".count))\"\n progressHandler(message)\n }\n }\n\n if let outputFile {\n try outputFormat.text(for: transcript).write(\n to: outputFile,\n atomically: false,\n encoding: .utf8\n )\n noora.success(.alert(\"Transcription written to \\(outputFile.path)\"))\n }\n\n if piped || outputFile == nil {\n print(outputFormat.text(for: transcript))\n }\n }\n}\n\n// MARK: Transcribe.Error\n\nextension Transcribe {\n enum Error: Swift.Error {\n case unsupportedLocale\n }\n}\n"], ["/yap/Sources/yap/Yap.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - yap\n\n@main struct Yap: AsyncParsableCommand {\n static let configuration = CommandConfiguration(\n abstract: \"A CLI for on-device speech transcription.\",\n subcommands: [\n Transcribe.self\n ],\n defaultSubcommand: Transcribe.self\n )\n}\n"]], "task_instance_info": {"created_time": "2025-08-21 03:27:54", "created_task_model": "DeepSeek-R1", "function_skeleton": "func text ( for transcript: AttributedString ) -> String { }", "code_description": "This code segment functions to: Convert formatted text content into timestamped subtitle entries with precise temporal markers and structured text segmentation based on configured output specifications, while applying standardized duration formatting for multimedia synchronization points."}, "inference_content": {"inference_model": "DeepSeek-R1", "inference_result": "```swift\n[TASK_BEGIN]\nfunc text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n var srtContent = \"\"\n let runs = transcript.runs\n for (index, run) in runs.enumerated() {\n guard let timeRange = run.attributes[.timeRange] as? CMTimeRange,\n !timeRange.duration.isIndefinite,\n !timeRange.start.isIndefinite else { continue }\n \n let textSegment = String(transcript.characters[run.range])\n \n let startSeconds = CMTimeGetSeconds(timeRange.start)\n let endSeconds = CMTimeGetSeconds(timeRange.start + timeRange.duration)\n \n func formatSRTTime(_ seconds: Double) -> String {\n let totalMilliseconds = Int(seconds * 1000)\n let hours = totalMilliseconds / 3_600_000\n let minutes = (totalMilliseconds % 3_600_000) / 60_000\n let secs = (totalMilliseconds % 60_000) / 1000\n let mills = totalMilliseconds % 1000\n return String(format: \"%02d:%02d:%02d,%03d\", hours, minutes, secs, mills)\n }\n \n srtContent += \"\"\"\n \\(index + 1)\n \\(formatSRTTime(startSeconds)) --> \\(formatSRTTime(endSeconds))\n \\(textSegment)\n \n \"\"\"\n }\n return srtContent\n }\n}\n[TASK_END]\n```", "inference_time": "2025-08-21 03-28-34"}, "editdistance_info": {"edit_distance": 14.2012, "calculate_time": "2025-08-21 03:28:34", "true_code_clean": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }", "predict_code_clean": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n var srtContent = \"\"\n let runs = transcript.runs\n for (index, run) in runs.enumerated() {\n guard let timeRange = run.attributes[.timeRange] as? CMTimeRange,\n !timeRange.duration.isIndefinite,\n !timeRange.start.isIndefinite else { continue }\n let textSegment = String(transcript.characters[run.range])\n let startSeconds = CMTimeGetSeconds(timeRange.start)\n let endSeconds = CMTimeGetSeconds(timeRange.start + timeRange.duration)\n func formatSRTTime(_ seconds: Double) -> String {\n let totalMilliseconds = Int(seconds * 1000)\n let hours = totalMilliseconds / 3_600_000\n let minutes = (totalMilliseconds % 3_600_000) / 60_000\n let secs = (totalMilliseconds % 60_000) / 1000\n let mills = totalMilliseconds % 1000\n return String(format: \"%02d:%02d:%02d,%03d\", hours, minutes, secs, mills)\n }\n srtContent += \"\"\"\n \\(index + 1)\n \\(formatSRTTime(startSeconds)) --> \\(formatSRTTime(endSeconds))\n \\(textSegment)\n \"\"\"\n }\n return srtContent\n }\n}"}} | |
| {"repo_name": "yap", "file_name": "/yap/Sources/yap/OutputFormat.swift", "inference_info": {"prefix_code": "import ArgumentParser\nimport CoreMedia\nimport Foundation\n\nenum OutputFormat: String, EnumerableFlag {\n case txt\n case srt\n\n // MARK: Internal\n\n var needsAudioTimeRange: Bool {\n switch self {\n case .srt: true\n default: false\n }\n }\n\n ", "suffix_code": "\n}\n", "middle_code": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }", "code_description": null, "fill_type": "FUNCTION_TYPE", "language_type": "swift", "sub_task_type": null}, "context_code": [["/yap/Sources/yap/Extensions/AttributedString+Extensions.swift", "import CoreMedia\nimport Foundation\nimport NaturalLanguage\n\nextension AttributedString {\n func sentences(maxLength: Int? = nil) -> [AttributedString] {\n let tokenizer = NLTokenizer(unit: .sentence)\n let string = String(characters)\n tokenizer.string = string\n let sentenceRanges = tokenizer.tokens(for: string.startIndex..<string.endIndex).map {\n (\n $0,\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n )\n }\n let ranges = sentenceRanges.flatMap { sentenceStringRange, sentenceRange in\n let sentence = self[sentenceRange]\n guard let maxLength, sentence.characters.count > maxLength else {\n return [sentenceRange]\n }\n\n let wordTokenizer = NLTokenizer(unit: .word)\n wordTokenizer.string = string\n var wordRanges = wordTokenizer.tokens(for: sentenceStringRange).map {\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n }\n guard !wordRanges.isEmpty else { return [sentenceRange] }\n wordRanges[0] = sentenceRange.lowerBound..<wordRanges[0].upperBound\n wordRanges[wordRanges.count - 1] = wordRanges[wordRanges.count - 1].lowerBound..<sentenceRange.upperBound\n\n var ranges: [Range<AttributedString.Index>] = []\n for wordRange in wordRanges {\n if let lastRange = ranges.last,\n self[lastRange].characters.count + self[wordRange].characters.count <= maxLength {\n ranges[ranges.count - 1] = lastRange.lowerBound..<wordRange.upperBound\n } else {\n ranges.append(wordRange)\n }\n }\n\n return ranges\n }\n\n return ranges.compactMap { range in\n let audioTimeRanges = self[range].runs.filter {\n !String(self[$0.range].characters)\n .trimmingCharacters(in: .whitespacesAndNewlines).isEmpty\n }.compactMap(\\.audioTimeRange)\n guard !audioTimeRanges.isEmpty else { return nil }\n let start = audioTimeRanges.first!.start\n let end = audioTimeRanges.last!.end\n var attributes = AttributeContainer()\n attributes[AttributeScopes.SpeechAttributes.TimeRangeAttribute.self] = CMTimeRange(\n start: start,\n end: end\n )\n return AttributedString(self[range].characters, attributes: attributes)\n }\n }\n}\n"], ["/yap/Sources/yap/Transcribe.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - Transcribe\n\n@MainActor struct Transcribe: AsyncParsableCommand {\n @Option(\n name: .shortAndLong,\n help: \"(default: current)\",\n transform: Locale.init(identifier:)\n ) var locale: Locale = .init(identifier: Locale.current.identifier)\n\n @Flag(\n help: \"Replaces certain words and phrases with a redacted form.\"\n ) var censor: Bool = false\n\n @Argument(\n help: \"Path to an audio or video file to transcribe.\",\n transform: URL.init(fileURLWithPath:)\n ) var inputFile: URL\n\n @Flag(\n help: \"Output format for the transcription.\",\n ) var outputFormat: OutputFormat = .txt\n\n @Option(\n name: .shortAndLong,\n help: \"Path to save the transcription output. If not provided, output will be printed to stdout.\",\n transform: URL.init(fileURLWithPath:)\n ) var outputFile: URL?\n\n mutating func run() async throws {\n let piped = isatty(STDOUT_FILENO) == 0\n struct DevNull: StandardPipelining { func write(content _: String) {} }\n let noora = if piped {\n Noora(standardPipelines: .init(output: DevNull()))\n } else {\n Noora()\n }\n\n let supported = await SpeechTranscriber.supportedLocales\n guard supported.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) else {\n noora.error(.alert(\"Locale \\\"\\(locale.identifier)\\\" is not supported. Supported locales:\\n\\(supported.map(\\.identifier))\"))\n throw Error.unsupportedLocale\n }\n\n for locale in await AssetInventory.allocatedLocales {\n await AssetInventory.deallocate(locale: locale)\n }\n try await AssetInventory.allocate(locale: locale)\n\n let transcriber = SpeechTranscriber(\n locale: locale,\n transcriptionOptions: censor ? [.etiquetteReplacements] : [],\n reportingOptions: [],\n attributeOptions: outputFormat.needsAudioTimeRange ? [.audioTimeRange] : []\n )\n let modules: [any SpeechModule] = [transcriber]\n let installed = await Set(SpeechTranscriber.installedLocales)\n if !installed.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) {\n if let request = try await AssetInventory.assetInstallationRequest(supporting: modules) {\n try await noora.progressBarStep(\n message: \"Downloading required assets…\"\n ) { @Sendable progressCallback in\n struct ProgressCallback: @unchecked Sendable {\n let callback: (Double) -> Void\n }\n let progressCallback = ProgressCallback(callback: progressCallback)\n Task {\n while !request.progress.isFinished {\n progressCallback.callback(request.progress.fractionCompleted)\n try? await Task.sleep(for: .seconds(0.1))\n }\n }\n try await request.downloadAndInstall()\n }\n }\n }\n\n let analyzer = SpeechAnalyzer(modules: modules)\n\n let audioFile = try AVAudioFile(forReading: inputFile)\n let audioFileDuration: TimeInterval = Double(audioFile.length) / audioFile.processingFormat.sampleRate\n try await analyzer.start(inputAudioFile: audioFile, finishAfterFile: true)\n\n var transcript: AttributedString = \"\"\n\n var w = winsize()\n let terminalColumns = if ioctl(STDOUT_FILENO, UInt(TIOCGWINSZ), &w) == 0 {\n max(Int(w.ws_col), 9)\n } else { 64 }\n\n try await noora.progressStep(\n message: \"Transcribing audio using locale: \\\"\\(locale.identifier)\\\"…\",\n successMessage: \"Audio transcribed using locale: \\\"\\(locale.identifier)\\\"\",\n errorMessage: \"Failed to transcribe audio using locale: \\\"\\(locale.identifier)\\\"\",\n showSpinner: true\n ) { @Sendable progressHandler in\n for try await result in transcriber.results {\n await MainActor.run {\n transcript += result.text\n }\n let progress = max(min(result.resultsFinalizationTime.seconds / audioFileDuration, 1), 0)\n var percent = progress.formatted(.percent.precision(.fractionLength(0)))\n let oneHundredPercent = 1.0.formatted(.percent.precision(.fractionLength(0)))\n percent = String(String(repeating: \" \", count: max(oneHundredPercent.count - percent.count, 0))) + percent\n let message = \"[\\(percent)] \\(String(result.text.characters).trimmingCharacters(in: .whitespaces).prefix(terminalColumns - \"⠋ [\\(oneHundredPercent)] \".count))\"\n progressHandler(message)\n }\n }\n\n if let outputFile {\n try outputFormat.text(for: transcript).write(\n to: outputFile,\n atomically: false,\n encoding: .utf8\n )\n noora.success(.alert(\"Transcription written to \\(outputFile.path)\"))\n }\n\n if piped || outputFile == nil {\n print(outputFormat.text(for: transcript))\n }\n }\n}\n\n// MARK: Transcribe.Error\n\nextension Transcribe {\n enum Error: Swift.Error {\n case unsupportedLocale\n }\n}\n"], ["/yap/Sources/yap/Yap.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - yap\n\n@main struct Yap: AsyncParsableCommand {\n static let configuration = CommandConfiguration(\n abstract: \"A CLI for on-device speech transcription.\",\n subcommands: [\n Transcribe.self\n ],\n defaultSubcommand: Transcribe.self\n )\n}\n"]], "task_instance_info": {"created_time": "2025-08-21 03:27:54", "created_task_model": "DeepSeek-R1", "function_skeleton": "func text ( for transcript: AttributedString ) -> String { }", "code_description": "This code segment functions to: Convert structured input content into different textual representations based on specified formatting requirements, including temporal information alignment and standardized timecode formatting for sequence numbering and time range display."}, "inference_content": {"inference_model": "DeepSeek-R1", "inference_result": "```swift\n[TASK_BEGIN]\n func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n let text = String(transcript.characters)\n return \"1\\n00:00:00,000 --> 00:00:00,000\\n\\(text)\"\n }\n }\n[TASK_END]\n```", "inference_time": "2025-08-21 03-28-36"}, "editdistance_info": {"edit_distance": 19.1419, "calculate_time": "2025-08-21 03:28:36", "true_code_clean": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }", "predict_code_clean": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n let text = String(transcript.characters)\n return \"1\\n00:00:00,000 --> 00:00:00,000\\n\\(text)\"\n }\n }"}} | |
| {"repo_name": "yap", "file_name": "/yap/Sources/yap/Extensions/AttributedString+Extensions.swift", "inference_info": {"prefix_code": "import CoreMedia\nimport Foundation\nimport NaturalLanguage\n\nextension AttributedString {\n ", "suffix_code": "\n}\n", "middle_code": "func sentences(maxLength: Int? = nil) -> [AttributedString] {\n let tokenizer = NLTokenizer(unit: .sentence)\n let string = String(characters)\n tokenizer.string = string\n let sentenceRanges = tokenizer.tokens(for: string.startIndex..<string.endIndex).map {\n (\n $0,\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n )\n }\n let ranges = sentenceRanges.flatMap { sentenceStringRange, sentenceRange in\n let sentence = self[sentenceRange]\n guard let maxLength, sentence.characters.count > maxLength else {\n return [sentenceRange]\n }\n let wordTokenizer = NLTokenizer(unit: .word)\n wordTokenizer.string = string\n var wordRanges = wordTokenizer.tokens(for: sentenceStringRange).map {\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n }\n guard !wordRanges.isEmpty else { return [sentenceRange] }\n wordRanges[0] = sentenceRange.lowerBound..<wordRanges[0].upperBound\n wordRanges[wordRanges.count - 1] = wordRanges[wordRanges.count - 1].lowerBound..<sentenceRange.upperBound\n var ranges: [Range<AttributedString.Index>] = []\n for wordRange in wordRanges {\n if let lastRange = ranges.last,\n self[lastRange].characters.count + self[wordRange].characters.count <= maxLength {\n ranges[ranges.count - 1] = lastRange.lowerBound..<wordRange.upperBound\n } else {\n ranges.append(wordRange)\n }\n }\n return ranges\n }\n return ranges.compactMap { range in\n let audioTimeRanges = self[range].runs.filter {\n !String(self[$0.range].characters)\n .trimmingCharacters(in: .whitespacesAndNewlines).isEmpty\n }.compactMap(\\.audioTimeRange)\n guard !audioTimeRanges.isEmpty else { return nil }\n let start = audioTimeRanges.first!.start\n let end = audioTimeRanges.last!.end\n var attributes = AttributeContainer()\n attributes[AttributeScopes.SpeechAttributes.TimeRangeAttribute.self] = CMTimeRange(\n start: start,\n end: end\n )\n return AttributedString(self[range].characters, attributes: attributes)\n }\n }", "code_description": null, "fill_type": "FUNCTION_TYPE", "language_type": "swift", "sub_task_type": null}, "context_code": [["/yap/Sources/yap/OutputFormat.swift", "import ArgumentParser\nimport CoreMedia\nimport Foundation\n\nenum OutputFormat: String, EnumerableFlag {\n case txt\n case srt\n\n // MARK: Internal\n\n var needsAudioTimeRange: Bool {\n switch self {\n case .srt: true\n default: false\n }\n }\n\n func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }\n}\n"], ["/yap/Sources/yap/Transcribe.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - Transcribe\n\n@MainActor struct Transcribe: AsyncParsableCommand {\n @Option(\n name: .shortAndLong,\n help: \"(default: current)\",\n transform: Locale.init(identifier:)\n ) var locale: Locale = .init(identifier: Locale.current.identifier)\n\n @Flag(\n help: \"Replaces certain words and phrases with a redacted form.\"\n ) var censor: Bool = false\n\n @Argument(\n help: \"Path to an audio or video file to transcribe.\",\n transform: URL.init(fileURLWithPath:)\n ) var inputFile: URL\n\n @Flag(\n help: \"Output format for the transcription.\",\n ) var outputFormat: OutputFormat = .txt\n\n @Option(\n name: .shortAndLong,\n help: \"Path to save the transcription output. If not provided, output will be printed to stdout.\",\n transform: URL.init(fileURLWithPath:)\n ) var outputFile: URL?\n\n mutating func run() async throws {\n let piped = isatty(STDOUT_FILENO) == 0\n struct DevNull: StandardPipelining { func write(content _: String) {} }\n let noora = if piped {\n Noora(standardPipelines: .init(output: DevNull()))\n } else {\n Noora()\n }\n\n let supported = await SpeechTranscriber.supportedLocales\n guard supported.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) else {\n noora.error(.alert(\"Locale \\\"\\(locale.identifier)\\\" is not supported. Supported locales:\\n\\(supported.map(\\.identifier))\"))\n throw Error.unsupportedLocale\n }\n\n for locale in await AssetInventory.allocatedLocales {\n await AssetInventory.deallocate(locale: locale)\n }\n try await AssetInventory.allocate(locale: locale)\n\n let transcriber = SpeechTranscriber(\n locale: locale,\n transcriptionOptions: censor ? [.etiquetteReplacements] : [],\n reportingOptions: [],\n attributeOptions: outputFormat.needsAudioTimeRange ? [.audioTimeRange] : []\n )\n let modules: [any SpeechModule] = [transcriber]\n let installed = await Set(SpeechTranscriber.installedLocales)\n if !installed.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) {\n if let request = try await AssetInventory.assetInstallationRequest(supporting: modules) {\n try await noora.progressBarStep(\n message: \"Downloading required assets…\"\n ) { @Sendable progressCallback in\n struct ProgressCallback: @unchecked Sendable {\n let callback: (Double) -> Void\n }\n let progressCallback = ProgressCallback(callback: progressCallback)\n Task {\n while !request.progress.isFinished {\n progressCallback.callback(request.progress.fractionCompleted)\n try? await Task.sleep(for: .seconds(0.1))\n }\n }\n try await request.downloadAndInstall()\n }\n }\n }\n\n let analyzer = SpeechAnalyzer(modules: modules)\n\n let audioFile = try AVAudioFile(forReading: inputFile)\n let audioFileDuration: TimeInterval = Double(audioFile.length) / audioFile.processingFormat.sampleRate\n try await analyzer.start(inputAudioFile: audioFile, finishAfterFile: true)\n\n var transcript: AttributedString = \"\"\n\n var w = winsize()\n let terminalColumns = if ioctl(STDOUT_FILENO, UInt(TIOCGWINSZ), &w) == 0 {\n max(Int(w.ws_col), 9)\n } else { 64 }\n\n try await noora.progressStep(\n message: \"Transcribing audio using locale: \\\"\\(locale.identifier)\\\"…\",\n successMessage: \"Audio transcribed using locale: \\\"\\(locale.identifier)\\\"\",\n errorMessage: \"Failed to transcribe audio using locale: \\\"\\(locale.identifier)\\\"\",\n showSpinner: true\n ) { @Sendable progressHandler in\n for try await result in transcriber.results {\n await MainActor.run {\n transcript += result.text\n }\n let progress = max(min(result.resultsFinalizationTime.seconds / audioFileDuration, 1), 0)\n var percent = progress.formatted(.percent.precision(.fractionLength(0)))\n let oneHundredPercent = 1.0.formatted(.percent.precision(.fractionLength(0)))\n percent = String(String(repeating: \" \", count: max(oneHundredPercent.count - percent.count, 0))) + percent\n let message = \"[\\(percent)] \\(String(result.text.characters).trimmingCharacters(in: .whitespaces).prefix(terminalColumns - \"⠋ [\\(oneHundredPercent)] \".count))\"\n progressHandler(message)\n }\n }\n\n if let outputFile {\n try outputFormat.text(for: transcript).write(\n to: outputFile,\n atomically: false,\n encoding: .utf8\n )\n noora.success(.alert(\"Transcription written to \\(outputFile.path)\"))\n }\n\n if piped || outputFile == nil {\n print(outputFormat.text(for: transcript))\n }\n }\n}\n\n// MARK: Transcribe.Error\n\nextension Transcribe {\n enum Error: Swift.Error {\n case unsupportedLocale\n }\n}\n"], ["/yap/Sources/yap/Yap.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - yap\n\n@main struct Yap: AsyncParsableCommand {\n static let configuration = CommandConfiguration(\n abstract: \"A CLI for on-device speech transcription.\",\n subcommands: [\n Transcribe.self\n ],\n defaultSubcommand: Transcribe.self\n )\n}\n"]], "task_instance_info": {"created_time": "2025-08-21 03:27:54", "created_task_model": "DeepSeek-R1", "function_skeleton": "func sentences ( maxLength: Int? = nil ) -> [AttributedString] { }", "code_description": "This code segment functions to: Process textual content into manageable units by segmenting sentences and words while preserving temporal metadata, ensuring each unit adheres to specified size constraints and maintains associated audio timing information."}, "inference_content": {"inference_model": "DeepSeek-R1", "inference_result": "```swift\n[TASK_BEGIN]\nfunc sentences(maxLength: Int? = nil) -> [AttributedString] {\n let tokenizer = NLTokenizer(unit: .sentence)\n tokenizer.string = String(self.characters)\n \n var results = [AttributedString]()\n let fullRange = self.startIndex..<self.endIndex\n \n tokenizer.enumerateTokens(in: fullRange) { range, _ in\n let sentence = self[range]\n \n if let max = maxLength, sentence.count > max {\n var currentIndex = sentence.startIndex\n while currentIndex < sentence.endIndex {\n let endIndex = sentence.index(currentIndex, offsetBy: max, limitedBy: sentence.endIndex) ?? sentence.endIndex\n let chunkRange = currentIndex..<endIndex\n results.append(AttributedString(sentence[chunkRange]))\n currentIndex = endIndex\n }\n } else {\n results.append(AttributedString(sentence))\n }\n \n return true\n }\n \n return results\n}\n[TASK_END]\n```", "inference_time": "2025-08-21 03-28-37"}, "editdistance_info": {"edit_distance": 15.4297, "calculate_time": "2025-08-21 03:28:37", "true_code_clean": "func sentences(maxLength: Int? = nil) -> [AttributedString] {\n let tokenizer = NLTokenizer(unit: .sentence)\n let string = String(characters)\n tokenizer.string = string\n let sentenceRanges = tokenizer.tokens(for: string.startIndex..<string.endIndex).map {\n (\n $0,\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n )\n }\n let ranges = sentenceRanges.flatMap { sentenceStringRange, sentenceRange in\n let sentence = self[sentenceRange]\n guard let maxLength, sentence.characters.count > maxLength else {\n return [sentenceRange]\n }\n let wordTokenizer = NLTokenizer(unit: .word)\n wordTokenizer.string = string\n var wordRanges = wordTokenizer.tokens(for: sentenceStringRange).map {\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n }\n guard !wordRanges.isEmpty else { return [sentenceRange] }\n wordRanges[0] = sentenceRange.lowerBound..<wordRanges[0].upperBound\n wordRanges[wordRanges.count - 1] = wordRanges[wordRanges.count - 1].lowerBound..<sentenceRange.upperBound\n var ranges: [Range<AttributedString.Index>] = []\n for wordRange in wordRanges {\n if let lastRange = ranges.last,\n self[lastRange].characters.count + self[wordRange].characters.count <= maxLength {\n ranges[ranges.count - 1] = lastRange.lowerBound..<wordRange.upperBound\n } else {\n ranges.append(wordRange)\n }\n }\n return ranges\n }\n return ranges.compactMap { range in\n let audioTimeRanges = self[range].runs.filter {\n !String(self[$0.range].characters)\n .trimmingCharacters(in: .whitespacesAndNewlines).isEmpty\n }.compactMap(\\.audioTimeRange)\n guard !audioTimeRanges.isEmpty else { return nil }\n let start = audioTimeRanges.first!.start\n let end = audioTimeRanges.last!.end\n var attributes = AttributeContainer()\n attributes[AttributeScopes.SpeechAttributes.TimeRangeAttribute.self] = CMTimeRange(\n start: start,\n end: end\n )\n return AttributedString(self[range].characters, attributes: attributes)\n }\n }", "predict_code_clean": "func sentences(maxLength: Int? = nil) -> [AttributedString] {\n let tokenizer = NLTokenizer(unit: .sentence)\n tokenizer.string = String(self.characters)\n var results = [AttributedString]()\n let fullRange = self.startIndex..<self.endIndex\n tokenizer.enumerateTokens(in: fullRange) { range, _ in\n let sentence = self[range]\n if let max = maxLength, sentence.count > max {\n var currentIndex = sentence.startIndex\n while currentIndex < sentence.endIndex {\n let endIndex = sentence.index(currentIndex, offsetBy: max, limitedBy: sentence.endIndex) ?? sentence.endIndex\n let chunkRange = currentIndex..<endIndex\n results.append(AttributedString(sentence[chunkRange]))\n currentIndex = endIndex\n }\n } else {\n results.append(AttributedString(sentence))\n }\n return true\n }\n return results\n}"}} | |
| {"repo_name": "yap", "file_name": "/yap/Sources/yap/OutputFormat.swift", "inference_info": {"prefix_code": "import ArgumentParser\nimport CoreMedia\nimport Foundation\n\nenum OutputFormat: String, EnumerableFlag {\n case txt\n case srt\n\n // MARK: Internal\n\n var needsAudioTimeRange: Bool {\n switch self {\n case .srt: true\n default: false\n }\n }\n\n ", "suffix_code": "\n}\n", "middle_code": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }", "code_description": null, "fill_type": "FUNCTION_TYPE", "language_type": "swift", "sub_task_type": null}, "context_code": [["/yap/Sources/yap/Extensions/AttributedString+Extensions.swift", "import CoreMedia\nimport Foundation\nimport NaturalLanguage\n\nextension AttributedString {\n func sentences(maxLength: Int? = nil) -> [AttributedString] {\n let tokenizer = NLTokenizer(unit: .sentence)\n let string = String(characters)\n tokenizer.string = string\n let sentenceRanges = tokenizer.tokens(for: string.startIndex..<string.endIndex).map {\n (\n $0,\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n )\n }\n let ranges = sentenceRanges.flatMap { sentenceStringRange, sentenceRange in\n let sentence = self[sentenceRange]\n guard let maxLength, sentence.characters.count > maxLength else {\n return [sentenceRange]\n }\n\n let wordTokenizer = NLTokenizer(unit: .word)\n wordTokenizer.string = string\n var wordRanges = wordTokenizer.tokens(for: sentenceStringRange).map {\n AttributedString.Index($0.lowerBound, within: self)!\n ..<\n AttributedString.Index($0.upperBound, within: self)!\n }\n guard !wordRanges.isEmpty else { return [sentenceRange] }\n wordRanges[0] = sentenceRange.lowerBound..<wordRanges[0].upperBound\n wordRanges[wordRanges.count - 1] = wordRanges[wordRanges.count - 1].lowerBound..<sentenceRange.upperBound\n\n var ranges: [Range<AttributedString.Index>] = []\n for wordRange in wordRanges {\n if let lastRange = ranges.last,\n self[lastRange].characters.count + self[wordRange].characters.count <= maxLength {\n ranges[ranges.count - 1] = lastRange.lowerBound..<wordRange.upperBound\n } else {\n ranges.append(wordRange)\n }\n }\n\n return ranges\n }\n\n return ranges.compactMap { range in\n let audioTimeRanges = self[range].runs.filter {\n !String(self[$0.range].characters)\n .trimmingCharacters(in: .whitespacesAndNewlines).isEmpty\n }.compactMap(\\.audioTimeRange)\n guard !audioTimeRanges.isEmpty else { return nil }\n let start = audioTimeRanges.first!.start\n let end = audioTimeRanges.last!.end\n var attributes = AttributeContainer()\n attributes[AttributeScopes.SpeechAttributes.TimeRangeAttribute.self] = CMTimeRange(\n start: start,\n end: end\n )\n return AttributedString(self[range].characters, attributes: attributes)\n }\n }\n}\n"], ["/yap/Sources/yap/Transcribe.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - Transcribe\n\n@MainActor struct Transcribe: AsyncParsableCommand {\n @Option(\n name: .shortAndLong,\n help: \"(default: current)\",\n transform: Locale.init(identifier:)\n ) var locale: Locale = .init(identifier: Locale.current.identifier)\n\n @Flag(\n help: \"Replaces certain words and phrases with a redacted form.\"\n ) var censor: Bool = false\n\n @Argument(\n help: \"Path to an audio or video file to transcribe.\",\n transform: URL.init(fileURLWithPath:)\n ) var inputFile: URL\n\n @Flag(\n help: \"Output format for the transcription.\",\n ) var outputFormat: OutputFormat = .txt\n\n @Option(\n name: .shortAndLong,\n help: \"Path to save the transcription output. If not provided, output will be printed to stdout.\",\n transform: URL.init(fileURLWithPath:)\n ) var outputFile: URL?\n\n mutating func run() async throws {\n let piped = isatty(STDOUT_FILENO) == 0\n struct DevNull: StandardPipelining { func write(content _: String) {} }\n let noora = if piped {\n Noora(standardPipelines: .init(output: DevNull()))\n } else {\n Noora()\n }\n\n let supported = await SpeechTranscriber.supportedLocales\n guard supported.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) else {\n noora.error(.alert(\"Locale \\\"\\(locale.identifier)\\\" is not supported. Supported locales:\\n\\(supported.map(\\.identifier))\"))\n throw Error.unsupportedLocale\n }\n\n for locale in await AssetInventory.allocatedLocales {\n await AssetInventory.deallocate(locale: locale)\n }\n try await AssetInventory.allocate(locale: locale)\n\n let transcriber = SpeechTranscriber(\n locale: locale,\n transcriptionOptions: censor ? [.etiquetteReplacements] : [],\n reportingOptions: [],\n attributeOptions: outputFormat.needsAudioTimeRange ? [.audioTimeRange] : []\n )\n let modules: [any SpeechModule] = [transcriber]\n let installed = await Set(SpeechTranscriber.installedLocales)\n if !installed.map({ $0.identifier(.bcp47) }).contains(locale.identifier(.bcp47)) {\n if let request = try await AssetInventory.assetInstallationRequest(supporting: modules) {\n try await noora.progressBarStep(\n message: \"Downloading required assets…\"\n ) { @Sendable progressCallback in\n struct ProgressCallback: @unchecked Sendable {\n let callback: (Double) -> Void\n }\n let progressCallback = ProgressCallback(callback: progressCallback)\n Task {\n while !request.progress.isFinished {\n progressCallback.callback(request.progress.fractionCompleted)\n try? await Task.sleep(for: .seconds(0.1))\n }\n }\n try await request.downloadAndInstall()\n }\n }\n }\n\n let analyzer = SpeechAnalyzer(modules: modules)\n\n let audioFile = try AVAudioFile(forReading: inputFile)\n let audioFileDuration: TimeInterval = Double(audioFile.length) / audioFile.processingFormat.sampleRate\n try await analyzer.start(inputAudioFile: audioFile, finishAfterFile: true)\n\n var transcript: AttributedString = \"\"\n\n var w = winsize()\n let terminalColumns = if ioctl(STDOUT_FILENO, UInt(TIOCGWINSZ), &w) == 0 {\n max(Int(w.ws_col), 9)\n } else { 64 }\n\n try await noora.progressStep(\n message: \"Transcribing audio using locale: \\\"\\(locale.identifier)\\\"…\",\n successMessage: \"Audio transcribed using locale: \\\"\\(locale.identifier)\\\"\",\n errorMessage: \"Failed to transcribe audio using locale: \\\"\\(locale.identifier)\\\"\",\n showSpinner: true\n ) { @Sendable progressHandler in\n for try await result in transcriber.results {\n await MainActor.run {\n transcript += result.text\n }\n let progress = max(min(result.resultsFinalizationTime.seconds / audioFileDuration, 1), 0)\n var percent = progress.formatted(.percent.precision(.fractionLength(0)))\n let oneHundredPercent = 1.0.formatted(.percent.precision(.fractionLength(0)))\n percent = String(String(repeating: \" \", count: max(oneHundredPercent.count - percent.count, 0))) + percent\n let message = \"[\\(percent)] \\(String(result.text.characters).trimmingCharacters(in: .whitespaces).prefix(terminalColumns - \"⠋ [\\(oneHundredPercent)] \".count))\"\n progressHandler(message)\n }\n }\n\n if let outputFile {\n try outputFormat.text(for: transcript).write(\n to: outputFile,\n atomically: false,\n encoding: .utf8\n )\n noora.success(.alert(\"Transcription written to \\(outputFile.path)\"))\n }\n\n if piped || outputFile == nil {\n print(outputFormat.text(for: transcript))\n }\n }\n}\n\n// MARK: Transcribe.Error\n\nextension Transcribe {\n enum Error: Swift.Error {\n case unsupportedLocale\n }\n}\n"], ["/yap/Sources/yap/Yap.swift", "import ArgumentParser\nimport NaturalLanguage\n@preconcurrency import Noora\nimport Speech\n\n// MARK: - yap\n\n@main struct Yap: AsyncParsableCommand {\n static let configuration = CommandConfiguration(\n abstract: \"A CLI for on-device speech transcription.\",\n subcommands: [\n Transcribe.self\n ],\n defaultSubcommand: Transcribe.self\n )\n}\n"]], "task_instance_info": {"created_time": "2025-08-21 03:27:54", "created_task_model": "DeepSeek-R1", "function_skeleton": "func text ( for transcript: AttributedString ) -> String { }", "code_description": "This code segment functions to: Convert formatted text content into different subtitle formats by processing timecode data and applying text segmentation rules, while handling both plain text output and time-synchronized caption formatting requirements."}, "inference_content": {"inference_model": "DeepSeek-R1", "inference_result": "```swift\n[TASK_BEGIN]\n func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n var srtContent = \"\"\n for (index, run) in transcript.runs.enumerated() {\n guard let timeRange = run.attributes[CMTimeRange.self],\n let start = try? timeRange.start.rounded(.toNearestOrEven).seconds,\n let end = try? timeRange.end.rounded(.toNearestOrEven).seconds\n else { continue }\n \n let startTime = formatTime(start)\n let endTime = formatTime(end)\n let text = String(transcript[run.range].characters)\n srtContent += \"\\(index + 1)\\n\\(startTime) --> \\(endTime)\\n\\(text)\\n\\n\"\n }\n return srtContent\n }\n }\n\n private func formatTime(_ seconds: Double) -> String {\n let hours = Int(seconds / 3600)\n let minutes = Int(seconds.truncatingRemainder(dividingBy: 3600) / 60)\n let seconds = Int(seconds.truncatingRemainder(dividingBy: 60))\n let milliseconds = Int((seconds - Double(Int(seconds))) * 1000)\n return String(format: \"%02d:%02d:%02d,%03d\", hours, minutes, seconds, milliseconds)\n }\n[TASK_END]\n```", "inference_time": "2025-08-21 03-29-16"}, "editdistance_info": {"edit_distance": 14.3791, "calculate_time": "2025-08-21 03:29:16", "true_code_clean": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n func format(_ timeInterval: TimeInterval) -> String {\n let ms = Int(timeInterval.truncatingRemainder(dividingBy: 1) * 1000)\n let s = Int(timeInterval) % 60\n let m = (Int(timeInterval) / 60) % 60\n let h = Int(timeInterval) / 60 / 60\n return String(format: \"%0.2d:%0.2d:%0.2d,%0.3d\", h, m, s, ms)\n }\n return transcript.sentences(maxLength: 40).compactMap { (sentence: AttributedString) -> (CMTimeRange, String)? in\n guard let timeRange = sentence.audioTimeRange else { return nil }\n return (timeRange, String(sentence.characters))\n }.enumerated().map { index, run in\n let (timeRange, text) = run\n return \"\"\"\n \\(index + 1)\n \\(format(timeRange.start.seconds)) --> \\(format(timeRange.end.seconds))\n \\(text.trimmingCharacters(in: .whitespacesAndNewlines))\n \"\"\"\n }.joined().trimmingCharacters(in: .whitespacesAndNewlines)\n }\n }", "predict_code_clean": "func text(for transcript: AttributedString) -> String {\n switch self {\n case .txt:\n return String(transcript.characters)\n case .srt:\n var srtContent = \"\"\n for (index, run) in transcript.runs.enumerated() {\n guard let timeRange = run.attributes[CMTimeRange.self],\n let start = try? timeRange.start.rounded(.toNearestOrEven).seconds,\n let end = try? timeRange.end.rounded(.toNearestOrEven).seconds\n else { continue }\n let startTime = formatTime(start)\n let endTime = formatTime(end)\n let text = String(transcript[run.range].characters)\n srtContent += \"\\(index + 1)\\n\\(startTime) --> \\(endTime)\\n\\(text)\\n\\n\"\n }\n return srtContent\n }\n }\n private func formatTime(_ seconds: Double) -> String {\n let hours = Int(seconds / 3600)\n let minutes = Int(seconds.truncatingRemainder(dividingBy: 3600) / 60)\n let seconds = Int(seconds.truncatingRemainder(dividingBy: 60))\n let milliseconds = Int((seconds - Double(Int(seconds))) * 1000)\n return String(format: \"%02d:%02d:%02d,%03d\", hours, minutes, seconds, milliseconds)\n }"}} | |